PPT - NTTAC!

Download Report

Transcript PPT - NTTAC!

Evaluating Your Program:
Developing Successful and Sustainable
Programs
OJJDP New Grantee Orientation
April 6–7, 2010
Washington, DC
CSR, Incorporated
Outline
•
•
•
•
•
•
What is Program Evaluation?
The Logic Model
Evaluation Questions
Types of Program Evaluation
Data Collection and Analysis
Evidence-Based Programs
Program Evaluation
• Objective measurement and systematic
analysis to determine the manner and extent to
which a program achieves its intended goal.
Definitions adapted from the OMB Circular no. A-11
Part 6, Section 200
http://www.ojjdp.ncjrs.gov/grantees/pm/glossary.html
What Is Program Evaluation?
• There are many ways to define program evaluation
• Evaluation can be defined as the systematic process of
collecting and analyzing data in order to:
– Determine whether program objectives have been achieved;
– Make a decision about the program;
– Improve the delivery of services.
Boulmetis, J., & Dutwin, P. (2002). The ABC’s of Evaluation.
San Francisco: Jossey Bass.
Why Evaluate Your Program?
• Satisfy funding agency’s grant requirements
• Document program accomplishments
• Justify program
• Improve program activities and services
Why Evaluate a Program?
• Evaluation results provide useful feedback to:
– Sponsors and Funders
– Program Staff
– Clients
– Donors
– The Larger Community
First Steps in Program Evaluation
• Construct a logic model
• Identify performance measures/outputs
• Identify outcomes
Logic Model: Definition
• “Tool used to visually describe the
linkages between program goals,
activities, and expected outcomes.”
Definition from: http://www.jrsainfo.org/jjec/resources/definitions.html
Logic Model
Conditions
(Needs or Problems)
What community need
will the project fill?
(e.g., youth need for
positive social outlets)
Activities
What services will
the project offer?
(e.g., tutoring,
after school sports,
mentoring)
Outputs
How will you measure
project services?
(e.g., number of
youth served)
Outcomes
How will you know that
the community need has
decreased? (e.g., youth
have improved attitudes,
increased self-esteem;
youth delinquency or
substance use rates have
decreased; youth school
attendance improved)
Developing a Framework
Conditions
(Needs or Problems)
Activities
Outputs
Outcomes
What community need
will the project fill?
(e.g., youth need for
positive social outlets)
What services will
the project offer?
(e.g., tutoring,
after school sports,
mentoring)
How will you measure
project services?
(e.g., number of
youth served)
How will you know that
the community need has
decreased? (e.g., youth
have improved attitudes,
increased self-esteem;
youth delinquency or
substance use rates have
decreased; youth school
attendance improved)
High rates of truancy
among elementary
students (K to 5th
grade)
After-school tutoring
1. Hours of tutoring
2. Number of youth
tutored
3. Number of support
sessions provided
1. Improved attitudes
toward school
2. Improved self-esteem
3. Improved school
attendance
Services of a family
support coordinator
Developing a Framework
ADVANCED MODEL
ASSUMPTIONS: Principles, evidence-based approaches, and theories that guide the program
Situation
Analysis
Problem
Identification
Priority
Setting
Mission
Goals
Advance
Model
INPUT
ACTIVITIES
Resources and
Contributions.
Tasks conducted
by grantee’s staff,
subcontractor, or
volunteers.
Activities are
directly linked to
outputs.
willOUTPUTS
go here
Products and
services
delivered.
OUTCOME
Changes in individuals, agencies, systems,
and communities. Outcomes may be intended
or unintended.
Initial
Learning
Awareness
Knowledge
Attitude
Skills
Opinions
Aspirations
Motivations
Intermediate
Action
Behavior
Practice
Policies
Social Action
DecisionMaking
ENVIRONMENT: External and contextual factors that influence the program
Sources: GAO-02-923 –Strategies for Assessing How information Dissemination
Contributes to Agency Goals. GAO-03-9 –Efforts to strengthen the link between
resources and results at the administration of children and families. GAO/GGD-00-10
Managing for Results: Strengthening Regulatory Agencies' Performance Management.
Ellen Taylor-Powell, 2000. “A logic model: A program performance framework.“
University of Wisconsin-Cooperative Extension Program Evaluation Conference.
IMPACTS
Long-Term
Conditions
Social
Economic
Civic
Environmental
Why Develop a Program Logic Model?
•
Clearly identifies program elements
— goals, inputs, activities, outputs, and outcomes
•
•
•
Shows how program elements fit together
Illustrates the logic of how your program
works
Helps specify what to measure in an
evaluation
Why Develop a Program Logic Model?
• Builds a common and shared understanding
of the program’s goals, structure, need for
resources, and expected results among
program stakeholders;
• Provides a valuable and concise tool when
presenting on the program to other groups;
• Provides a record of your program.
How to Construct a Program Logic Model
There are several ways to do so:
• Convene a meeting of your program’s stakeholders
– Develop the model as a group.
(Building the model helps build consensus and
support for your program.)
• Program staff develop the logic model
– Meet separately with various stakeholder groups to
elicit their perceptions and expectations.
Different stakeholder groups may hold different views
about the program’s activities and expected results.
Where to Begin in Constructing a Logic Model
•
A program logic model is a type of flowchart
– One may begin constructing the model at either end of
the chart.
•
Try to construct a program logic model as early as possible
during program planning.
•
A program may be up and running before a logic model is
developed.
•
A logic model is a living document
– Revisions may be necessary as the program changes.
Next Steps in Program Evaluation
Once you have the logic model:
• Decide on your evaluation questions
(They are related to the goals of your program)
• How effective is your program?
Moving From Goals to Evaluation Questions
Project:
After-school
recreation
program
designed to
promote selfesteem and
reduce
delinquency
Program
Goals:
Increase
youth
protective
factor of selfesteem.
Lower arrest
rates among
program
youth.
Evaluatio
n Goal:
Determine
whether youth
served
increase their
self-esteem
and decrease
delinquency
Evaluation
Questions:
Do youth who
complete the
program show
increased selfesteem and
decreased
likelihood of
delinquency?
Types of Program Evaluation
• Process Evaluation
–
–
–
–
How does the program work?
What services are provided?
Is the program implemented as planned?
What are the outputs?
What works or does not work in the program?
Types of Program Evaluation
• Outcome Evaluation
– Assess program effectiveness
– What are the results?
• Short term
• Long term
Next Steps in Program Evaluation
Choose the type of evaluation to conduct:
• Decide on the data to be collected to monitor
performance measures/outputs and outcomes
• Create instruments needed to collect data
• Decide how to analyze and use data
Data Collection
• Need up-front planning
• Need a sense of what you are trying to accomplish
• What data will you collect and why?
• What data sources are available and which
will you use?
Types of Data Collection Tools
Good for “What?” and
“How many” questions
Activity Logs/
Skill Sheets
Written documentation of participant's
attendance, achievement, or acquisition of
Document/Records
review
Review of written documents such as
performance ratings, program logs, tally
sheets, and other existing indicators
Good for “What?” and
“How many?” questions
Focus groups
Moderated discussions on a particular topic
or issue
Good for “What?”
“How?” and “Why?”
questions
Interviews
Data collection through oral conversations
Good for “What?” and
“Why?” questions
Observation
Watching people engaged in activities and
recording what occurs
Good for “How?”
“What?” and “How
many?” questions
Questionnaires
Written responses to clearly defined
questions
Good for “What?” and
“How many?” questions
skills
www.evaluationtools.org/tools_main.asp
Create Instruments
• Construct your own instruments
– Attendance sheets
– Activity logs
– Surveys for clients
• Use existing instruments
– Templates available online
– Modify to suit your particular program
Ways to Collect Data
•
Paper
̶ Internal forms/checklists
•
•
Electronic
̶ Excel, Access, SPSS
Online
̶ Web-based data collection tools
̶ OJJDP’s Data Collection & Technical Assistance
Tool (DCTAT)
Analyze and Use Data
• Analyze Data
– How will the data be analyzed?
– Who will be responsible for data analysis?
• Use Data
– How will the data be used?
– Will the data be made available in reports?
– Who will receive the reports?
Evidence-Based Programs
• A program is evidence-based if:
– evaluation research shows that it produces the
expected positive results;
– the results can be attributed to the program itself, not
to other factors or events;
– the evaluation is peer-reviewed by experts in the
field;
– the program is “endorsed” by a federal agency or
respected research organization and included in its
list of effective programs.
From: Evidence-based Programs: An Overview
http://www.uwex.edu/ces/flp/families/whatworks_06.pdf
The Importance of Evidence-Based Practices
• Government-wide move toward accountability
• Government Performance and Results Act
– Shift from accountability for process to accountability
for results
– Programs must show effectiveness to justify funding
• President’s Management Agenda
• Evidence-based practices are cost effective
– It is important to provide programs that work
Program Evaluation: Success and Sustainability
• Use data to create interest and support for
program continuation
– From funders
– From the community
• Program evaluation helps identify the most
successful components of your program
– Maintain effective staff, procedures, or activities
Conclusion
• Program evaluation is an ongoing process
• Start with a logic model to identify the main
elements of the program
• Choose the type of evaluation needed
• Plan data collection and analysis activities in
advance
• Use program evaluation results to justify and
market the program
Resources
Boulmetis, J., & Dutwin, P. (2002). The ABC’s of Evaluation.
San Francisco: Jossey Bass.
The Program Manager’s Guide to Evaluation
Office of Planning, Research and Evaluation
Administration for Children and Families
U.S. Department of Health and Human Services
http://www.acf.hhs.gov/programs/opre/other_resrch/pm_guide_eval
Planning a Program Evaluation
University of Wisconsin Cooperative Extension
http://learningstore.uwex.edu/assets/pdfs/G3658-1.pdf
The Global Social Change Research Project
http://gsociology.icaap.org/methods/evaluationbeginnersguide.pdf
Contact Information
Monica Robbers, Ph.D.
Patricia San Antonio, Ph.D.
CSR, Incorporated
703.312.5220
[email protected]
[email protected]