Evaluation 101: An Introduction for New Evaluation Practitioners

Similar documents
Executive Summary. Abraxas Naperville Bridge. Eileen Roberts, Program Manager th St Woodridge, IL

K-12 Academic Intervention Plan. Academic Intervention Services (AIS) & Response to Intervention (RtI)

Cooking Matters at the Store Evaluation: Executive Summary

The Teenage Brain and Making Responsible Decisions About Sex

RtI: Changing the Role of the IAT

Your Guide to. Whole-School REFORM PIVOT PLAN. Strengthening Schools, Families & Communities

Strategic Planning for Retaining Women in Undergraduate Computing

Social Justice Practicum (SJP) Description

Multi Method Approaches to Monitoring Data Quality

DESIGNPRINCIPLES RUBRIC 3.0

Trends & Issues Report

Triple P Ontario Network Peaks and Valleys of Implementation HFCC Feb. 4, 2016

Newburgh Enlarged City School District Academic. Academic Intervention Services Plan

State Parental Involvement Plan

Volunteer State Community College Strategic Plan,

Evaluation Off Off On On

Strategic Practice: Career Practitioner Case Study

Port Jefferson Union Free School District. Response to Intervention (RtI) and Academic Intervention Services (AIS) PLAN

CORRELATION FLORIDA DEPARTMENT OF EDUCATION INSTRUCTIONAL MATERIALS CORRELATION COURSE STANDARDS / BENCHMARKS. 1 of 16

Department of Research & Program Evaluation (DRPE) Office of Accountability. Requests for Flexibility Evaluation Approach APPENDIX A

Early Warning System Implementation Guide

THE FIELD LEARNING PLAN

MSW POLICY, PLANNING & ADMINISTRATION (PP&A) CONCENTRATION

Study Abroad: Planning and Development, Successes and Challenges

Process Evaluations for a Multisite Nutrition Education Program

University of Arkansas at Little Rock Graduate Social Work Program Course Outline Spring 2014

(2) GRANT FOR RESIDENTIAL AND REINTEGRATION SERVICES.

OUCH! That Stereotype Hurts Cultural Competence & Linguistic Training Summary of Evaluation Results June 30, 2014

CONCEPT MAPS AS A DEVICE FOR LEARNING DATABASE CONCEPTS

Person Centered Positive Behavior Support Plan (PC PBS) Report Scoring Criteria & Checklist (Rev ) P. 1 of 8

What is PDE? Research Report. Paul Nichols

Massachusetts Juvenile Justice Education Case Study Results

This Access Agreement is for only, to align with the WPSA and in light of the Browne Review.

Executive Summary. Laurel County School District. Dr. Doug Bennett, Superintendent 718 N Main St London, KY

Number of students enrolled in the program in Fall, 2011: 20. Faculty member completing template: Molly Dugan (Date: 1/26/2012)

Success Factors for Creativity Workshops in RE

The One Minute Preceptor: 5 Microskills for One-On-One Teaching

Denver Public Schools

The feasibility, delivery and cost effectiveness of drink driving interventions: A qualitative analysis of professional stakeholders

Business. Pearson BTEC Level 1 Introductory in. Specification

COMMUNICATION PLAN. We believe that all individuals are valuable and worthy of respect.

OCR LEVEL 3 CAMBRIDGE TECHNICAL

STANDARDS AND RUBRICS FOR SCHOOL IMPROVEMENT 2005 REVISED EDITION

Welcome to the session on ACCUPLACER Policy Development. This session will touch upon common policy decisions an institution may encounter during the

NC Education Oversight Committee Meeting

Long Beach Unified School District

Strategic Plan Update Year 3 November 1, 2013

ESL Summer Camp: June 18 July 27, 2012 Homestay Application (Please answer all questions completely)

Tools to SUPPORT IMPLEMENTATION OF a monitoring system for regularly scheduled series

Assessment. the international training and education center on hiv. Continued on page 4

Illinois WIC Program Nutrition Practice Standards (NPS) Effective Secondary Education May 2013

Youth & Family Services Counseling Center

BSP !!! Trainer s Manual. Sheldon Loman, Ph.D. Portland State University. M. Kathleen Strickland-Cohen, Ph.D. University of Oregon

Short Term Action Plan (STAP)

PROPOSED MERGER - RESPONSE TO PUBLIC CONSULTATION

INDEPENDENT STUDY PROGRAM

TRANSNATIONAL TEACHING TEAMS INDUCTION PROGRAM OUTLINE FOR COURSE / UNIT COORDINATORS

The whole school approach and pastoral care

VIEW: An Assessment of Problem Solving Style

Tentative School Practicum/Internship Guide Subject to Change

Upward Bound Program

Modified Systematic Approach to Answering Questions J A M I L A H A L S A I D A N, M S C.

MARKETING FOR THE BOP WORKSHOP

School Health Survey, Texas Education Agency

Pierce County Schools. Pierce Truancy Reduction Protocol. Dr. Joy B. Williams Superintendent

School Health Survey, Texas Education Agency

Department of Social Work Master of Social Work Program

Educating Georgia s Future gadoe.org. Richard Woods, Georgia s School Superintendent. Richard Woods, Georgia s School Superintendent. gadoe.

ADDENDUM 2016 Template - Turnaround Option Plan (TOP) - Phases 1 and 2 St. Lucie Public Schools

Goal #1 Promote Excellence and Expand Current Graduate and Undergraduate Programs within CHHS

Executive Summary. Lava Heights Academy. Ms. Joette Hayden, Principal 730 Spring Dr. Toquerville, UT 84774

OVERVIEW OF CURRICULUM-BASED MEASUREMENT AS A GENERAL OUTCOME MEASURE

Post Test Attendance Record for online program and evaluation (2 pages) Complete the payment portion of the Attendance Record and enclose payment

This Access Agreement is for only, to align with the WPSA and in light of the Browne Review.

Running Head GAPSS PART A 1

Getting Results Continuous Improvement Plan

EDUC 2020: FOUNDATIONS OF MULTICULTURAL EDUCATION Spring 2011

Global School-based Student Health Survey (GSHS) and Global School Health Policy and Practices Survey (SHPPS): GSHS

Milton Public Schools Special Education Programs & Supports

Iowa School District Profiles. Le Mars

School-Wide Restorative Practices: Step by Step

INTRODUCTION TO PSYCHOLOGY

Prince2 Foundation and Practitioner Training Exam Preparation

EXPERIENCE UGA Outstanding Process Improvement: Increase Service to Students

A Framework for Safe and Successful Schools

MONTPELLIER FRENCH COURSE YOUTH APPLICATION FORM 2016

Guide for Fieldwork Educators

Upward Bound Math & Science Program

SSIS SEL Edition Overview Fall 2017

Lesson M4. page 1 of 2

Executive Summary. Walker County Board of Education. Dr. Jason Adkins, Superintendent 1710 Alabama Avenue Jasper, AL 35501

Examining the Structure of a Multidisciplinary Engineering Capstone Design Program

5 Early years providers

PRESENTED BY EDLY: FOR THE LOVE OF ABILITY

H2020 Marie Skłodowska Curie Innovative Training Networks Informal guidelines for the Mid-Term Meeting

RESIDENCE DON APPLICATION

Post-16 transport to education and training. Statutory guidance for local authorities

PERFORMING ARTS. Unit 2 Proposal for a commissioning brief Suite. Cambridge TECHNICALS LEVEL 3. L/507/6467 Guided learning hours: 60

Advances in Assessment The Wright Institute*

DICE - Final Report. Project Information Project Acronym DICE Project Title

Transcription:

Evaluation 101: An Introduction for New Evaluation Practitioners AEA/CDC Summer Evaluation Institute 2008 Introductions & Expectations Introduce yourself to someone at your table using these three basic topics Name, rank, serial number Something you love about the summer One expectation you have for today s session Be prepared to introduce your new friend to the group Evaluation 101 2 1

The Agenda 1. Welcome & Introductions 2. Establishing the Context 3. Basic Types of Evaluation 4. Descriptions and Examples of Each 5. Practice Developing Evaluations Evaluation 101 3 Types of Evaluation the Activities Formative evaluates a program and its process during development (formation) Process evaluates the process fidelity, implementation of the program compared to the design the Effect Outcome evaluates effectiveness in terms of programmatic outcomes Impact evaluates effect on community and other institutions Evaluation 101 4 2

Formative Evaluation What, Why, When Helps to identify or clarify the needs the new program is meant to address Helps identify gaps in services Tests initial reaction to the program design Used to pre-test a design before full implementation Sample Questions What is the most efficient way to recruit participants? What types of program activities are desired? What are the preferences of consumers? Evaluation 101 5 Formative Evaluation - examples Mass Mailing Should Land s End buy my address from Sears? STD testing program is planning to include urine based testing for female clients because it is less intrusive than pelvic exams. The formative evaluation results show that some do prefer the urine test because it s quick. But many don t believe their test is complete or that results are really valid. Evaluation 101 6 3

Designing a Formative Evaluation 1. Design Review Does the program include elements to address a particular need or client deficit? Does the program design match the intended client? 2. Expert Review Has the content or design been validated by experts or other research? Is the design consistent with current best practices in the field? 3. Client/Customer Review Is the message/program/service clearly understood by clients? What effects do program delivery have on program receipt? Evaluation 101 7 Process Evaluation What, Why, When looks at what activities, services or intervention is being implemented Accountability - Determine alignment with program s original design or purpose; for monitoring Program improvement - mid-course corrections, changes in outreach, recruitment, or data collection Replication clarify the ingredients before replicating or taking to scale Sample Questions Who is the intended target population of the program? Which elements of the program have actually been implemented? What barriers did clients experience in accessing the services? Evaluation 101 8 4

Process Evaluation - examples Bath Time We re done! We re ready for bed. But what really happened? Prenatal Teen Parent Education classes program is funded through a state health department grant and is required to use a particular curriculum. The new curriculum is being integrated into a program that is working with first time and second time teen mothers through the YWCA The process evaluation Clarify all of the services or interventions that are being implemented. See how well the instructor is following the curriculum How do other services influence uptake of information Evaluation 101 9 Designing a Process Evaluation 1. Determine purpose 2. Develop evaluation questions 3. Collect credible (quantifiable) evidence 4. Analyze data & justify conclusions 5. Report findings Evaluation 101 10 5

Designing a Process Evaluation 1. Determine purpose All programs new or existing have some purpose, concept or theory behind why they exist May require developing a logic model May be dictated by grant (often the case for government funding) Evaluation 101 11 Designing a Process Evaluation 1. Determine purpose 2. Develop evaluation questions Reach, Coverage - relates to the target population (characteristics, proportions served, outreach efforts) Dose, Duration relates to services or intervention (what services, how often, by who, cost) Context relates to other factors influencing how program was implemented (neighborhood, additional services) Fidelity relates to how well adhered to plan Evaluation 101 12 6

Designing a Process Evaluation 1. Determine purpose 2. Develop evaluation questions 3. Collect credible (quantifiable) evidence (Examples) Client demographics age, race, gender, socioeconomic status Client s prior status or behavior previous alcohol abuse, exercise, frequency of reading to their child Client outreach method of contact, mode of transportation Staff demographics, training, turnover rate Program intervention number of training sessions, number of condoms distributed, frequency and attendance at services Evaluation 101 13 Activity 1 Formative & Process Evaluations In small groups, review the scenario provided 1. Develop questions you would ask if conducting this evaluation 2. Develop a list of possible data points or evidence you might need to answer those questions Evaluation 101 14 7

Outcome Evaluation What, Why, When Measures the effect on clients, a population, or the community - changes in knowledge, attitude or behavior Improve the service delivery of the program by focusing on key tasks; Identify effective practices within the program Usually conducted after program has been implemented for enough time to plausibly anticipate results Sample Questions Are participants more knowledgeable about the subject after their training? Has there been a change in behavior (decrease in teen smoking) since the intervention began? Evaluation 101 15 Impact Evaluation What, Why, When Measures the effect on clients, a population, or the community Changes in knowledge attitude or behavior or condition Very similar to outcomes evaluation Sample Questions What is the effect of the program on the long term condition of a group or population? What is the collective affect of similar programs? How have these programs affected the system of services related to this need? Evaluation 101 16 8

Outcome Evaluation - examples Dinner If we re all still hungry was it a success? GED prep & Job readiness program is gets county funding and money from various other sources. It s program has two core components a 6-week GED preparation class and a 6- week job readiness program. Participants usually attend both either on same day or different days. The outcome evaluation looks at how many actually pass the GED test and how many ultimately get a job. It s not enough to look at program attendance or participant effort (i.e. creating a resume) Evaluation 101 17 Designing an Outcome Evaluation 1. Develop client outcome-based logic model 2. Identify clearly linked indicators 3. Collect credible (quantifiable) evidence 4. Analyze data & justify conclusions 5. Report findings Evaluation 101 18 9

Key Components of a Logic Model Inputs - Resources dedicated to or consumed by the program(s) within an agency and constraints on the agency Activities - What the agency does with the inputs to fulfill its mission the program services Outputs - The direct products of agency services the results of the process Outcomes - Benefits or changes to individuals during or after participating in program activities Evaluation 101 19 Logic Model INPUTS ACTIVITIES OUTPUTS OUTCOMES Evaluation 101 20 10

Layers of Outcomes Initial Initial - The most immediate benefits or changes participants experience and the ones most influenced by the program s outputs. Changes in knowledge, attitudes, or skills Intermediate LONG TERM Intermediate - Link a program s initial outcomes to the longer-term outcomes it desires for participants. Changes in behavior that result from the participants new knowledge, attitudes, or skills Long Term - The ultimate outcome that a program desires to achieve for its participants. Evaluation 101 21 Layers of Outcomes Initial Intermediate LONG TERM Teen Mother Parenting Education Program Teen mothers are knowledgeable of pre-natal nutrition and health guidelines. Teen mothers follow proper nutrition and health guidelines. Teen mothers deliver healthy babies. Evaluation 101 22 11

Sample Outcomes Outcomes should be linked in a logical fashion. If X happens then, Y will happen Program Type Parenting Education Initial Parents learn what their child is capable of doing age appropriate expectations Outcomes Intermediate Parents stay active with their child's education and learning Long Term Children are ready for school Emergency Financial Assistance Household needs are assessed and overdue rent is paid Households are provided with information on additional resources Households are stable Mental Health screenings Individuals are assessed for mental health issues & learn signs of mental illness Individuals are provided with referrals for treatment Evaluation 101 23 Sample Logic Model Inputs 2 certified teachers 1 PT MSW $50,000 from county $5,000 from UWMA Partnership with 3 local principals Free space in East town middle school Activities Daily homework assistance Nationally proven reading activities weekly Organized recreational activities Cultural/arts activities weekly Giving Kids a Chance After School Program (for middle school age children) Outputs Avg. daily attendance 45 3000 hours of homework time each quarter 2 organized intramural soccer teams in spring 25 hours of music class each quarter Initial Students enjoy learning Improved relationships with peers and adults Students complete homework each week Outcomes Intermediate Improved school attendance Improved grades in core subjects Long Term Maintain B average in core subjects Evaluation 101 24 12

Scoring Logic Models 1. In the packet are two logic models 2. Find a partner 3. Review the logic models and use the handout provided to create a score 4. Be sure to discuss your reasoning and if appropriate document ways that the logic model could be improved. Evaluation 101 25 Road Trip How do you know if the kids are behaving? Evaluation 101 26 13

Outcome Indicators Indicators Are the specific information that track a program s success. It s how you know something changed Traits of an Effective Indicator Measurable Meaningful Manageable As unbiased as possible Sensitive to change Acceptable to stakeholders Clear Evaluation 101 27 Outcome Indicators - Example Outcome Initial Parents learn what their children are capable of doing Intermediate Parents participate in their child s education Long Term Children are ready for school Indicator # of parents that demonstrate increased knowledge of child development through prepost test on 5 key issues after attending workshops # of parents that attend at least one school based event in addition to parent-teacher conferences # of children that are developmentally ready based on standardized child development assessment tool Evaluation 101 28 14

Steps to Writing a Good Indicator Identify exactly who is hoped to benefit (WHO?) Identify specific, observable change or accomplishment (WHAT?) Determine when the outcome is expected to occur (BY WHEN?) Indicator Example: WHO # of parents WHAT that demonstrate increased knowledge of child development through pre-post test on 5 key issues BY WHEN after attending workshops Evaluation 101 29 Targets & Benchmarks Target a numerical objective for a program s level of achievement on an indicator a projection Benchmark performance data used for comparison past agency data or industry standard Evaluation 101 30 15

Targets & Benchmarks example Outcome Initial Parents learn what their children are capable of doing Indicator # of parents that demonstrate increased knowledge of child development through pre-post test on 5 key issues after attending workshops Target 200 Benchmark or Target explanation Generally we see 200-300 parents each year. We believe that 90% of parents will show some improvement on the pre-post test. Last year 95% of parents showed increased knowledge Evaluation 101 31 Data Source & Data Collection Methods Data Source answers the question: Who or where will I get the information from? Data Collection Method answers the questions: What is the tool or method for collecting the data? How is the tool administered? How often is information collected? Evaluation 101 32 16

Data Source Pros & Cons Data Source Example Advantages Disadvantages Program Records (yours or others) Report cards, completion certificates, referrals Available Accessible Value of data depends on how carefully it was recorded Specific Individuals or Trained Observers Teachers report on student behavior, case manager, client Provides 1 st hand account Can be biased by interpretation or perceived pressure Mechanical Measurements Blood test, scale Relatively objective Findings are affected by accuracy of device Evaluation 101 33 Data Collection Methods Data Collection Method provides a description of the process for collecting the information Example A Annual review of program records of referrals sent for housing subsidy Example B Caseworkers rate the family each month during home visit Example C Tool = Self-Administered Questionnaire Distribution = sent via mail with stamped return envelope Frequency = sent 90 days after completion of program Evaluation 101 34 17

Data Collection example Outcome Initial Parents increase knowledge Indicator # of parents that demonstrate increased knowledge of child development through pre-post test on 5 key issues after attending workshops Target 200 Data Source The parents who participate in at least 2 sessions Data collection methods Written or online survey that is distributed at 1 st class and again at last class. Parents who did not complete both tests are not included in final results. Evaluation 101 35 Review & Closing Review expectations Questions, Comments Thank You Evaluation 101 36 18