Logic Model Development

advertisement
Measuring Program Results
for
Abstinence Education Grantees
Vijaya ChannahSorah, Ph.D.
Independent Consultant, Results Management
RESULTS
•
What do we mean by measuring results?
> Measuring Outputs and Outcomes
> Program Evaluation
> Logic Modeling (as a tool)
•
The purpose is to improve program outcomes and to
achieve our vision . . . to give our youth knowledge
and skills to live a rich, grounded life . . .
Page 2
Why are Results Important?
•
Enable you to know what works well (in what
populations, under which conditions)
•
Enable you to take corrective actions
•
Other states/projects can implement effective
program elements and test them in their own
area
Page 3
TRAINING APPROACH
Walk systematically through:
•
•
•
•
Performance Measurement (outputs,
outcomes, efficiency, etc.)
Program Evaluation
Logic Modeling
Tying Everything Together
Page 4
PERFORMANCE MEASUREMENT
•
Definitions: outputs, outcomes, etc.
•
Setting targets
•
How different from research / program
evaluation
Page 5
Performance Measurement:
Definitions
•
Outcomes: ultimate purpose of the
program
•
Outputs: intermediate results of
activities
•
Activities: your processes
•
Inputs: resources
•
Efficiency: outcomes or outputs
over costs (usually)
Page 6
Performance Measurement:
Definitions / Examples
Outcome example:
• Decrease the rate of births to unmarried teenage girls ages 15
to 19. (2002 baseline was 35.4%)
Output example:
• Number of teachers trained in abstinence education (by 200X).
• Activities: Developing curriculum, conducting training, etc.
• Inputs: Funds, x number of abstinence experts, ideas, etc.
•
Efficiency: Number of teachers trained over training dollars (a
PROCESS Efficiency Measure).
Page 7
Challenges in Performance Measurement
• Defining “efficiency” for human services
programs (e.g., abstinence education)
• Defining outcomes (e.g., family violence
prevention)
• Timely, reliable data
Page 8
Performance Measurement / Program
Evaluation: definitions
Difference between performance measurement
& program evaluation
Performance measurement shows:
• Trends over time
• Comparison between actual and desired results/outcomes
Program evaluation shows:
• Outcomes relative to what they would have been in absence of the
program
• Program’s causal contribution to the observed outcome
Page 9
PROGRAM EVALUATION
 Definition:
Program evaluations are systematic, scientifically-based
studies conducted to assess the impact of a program
A program evaluation typically applies scientific techniques to
determine how confident we can be that a particular
OUTCOME was CAUSED BY the intervention(s)
Evaluations examine achievement of program objectives in the
context of environmental and external factors (and takes them
into account)
Page 10
Program Evaluation
Basic Approaches:
 Randomized controlled trials
________________
O1
X
O2
-----------O1
O2
________________
Random assignment to control and experimental
groups
Page 11
Program Evaluation
Basic Approaches:
 Quasi experimental design
________________
O1
X
O2
-----------O1
O2
________________
No random assignment to control and experimental
groups
Page 12
Program Evaluation
Basic Approaches:
 Longitudinal quasi experimental design
_________________________________
O1
X
O2
X
O3
X
O4
_________________________________
Observations and treatments (interventions)
over time
Page 13
Program Evaluation
Basic Approaches:
 Single group pre- post-design
_______________
O1
X
O2
________________
No comparison/control group. Not recommended.
Conclusions about causality will be uncertain.
Page 14
Program Evaluation
Key things to look for (and discuss
with evaluator):
• Type and rigor of study design
• Hypotheses addressed?
Page 15
Program Evaluation
Key things to look for continued…
• Timing between pre- and post-tests (if
applicable)
• Data collected & methods of collection
• Frequency of data collection
• Demographics, external factors collected?
• Internal/external validity of study
Page 15 - A
Program Evaluation
Using milestones/interim results to take
corrective actions
• Ensure evaluation, data gathering, and outcomes
measurement allow for interim reporting
• Find out about the data lag (and implications)
• Determine how interim results can be interpreted (will
you know what to change about the program
elements/activities?)
Page 16
Program Evaluation
Reminders:
• Consider evaluating relatively small chunks of
program aspects (you may not be able to
conduct an evaluation of the entire project)
• Build use of interim results into your workplan
Page 17
Program Evaluation
Reminders (continued):
• Tap into partnerships: are existing data available (or
is information available) that might be useful?
• Check in “early and often” with 3rd party evaluator ~
stay in close contact
• Design the formats for evaluation reports (interim
and final) at the start
Page 18
Program Evaluation
Awareness of analysis techniques:
•
Multiple regression
•
Non-response analysis
•
Correlation (is NOT causality!)
Page 19
Program Evaluation
Feedback loop
•
Learn from the performance and program
evaluation process and results
•
Incorporate program evaluation into the logic
model (discussed next)
Page 20
LOGIC MODELS
A diagram, chart, or picture of all the major
elements of your entire project
The idea is to make it user-friendly to you and
your group, and applicable to your purpose
You will get basic building blocks here, and can
then tailor the logic model to your needs and
preferred ways of thinking
Page 21
Overview of Logic Models
• A logic model tracks how we get from our
challenges to our solutions and desired long
term outcomes …
CHALLENGE
INPUTS
ACTIVITIES
DEMOGRAPHICS
EXT. FACTORS
CONSTRAINTS
ASSUMPTIONS
OUTPUTS
OUTCOMES
LONG TERM
OUTCOMES
(GOALS)
Logic Model Defined
What is a Logic Model?
• A graphic representation of a program.
• It shows what the program is designed to
accomplish, including the services it
delivers, expected results of those services,
and the linkages between services and
program goals.
Page 23
Logic Models
…can go both ways
X ---------------------------------------------Y
(e.g., how can we make a raft float?)
X <- --------------------------------------------Y
(e.g., why did the raft sink?)
Page 24
Logic Model Uses
Use for program:
• Design
• Budgeting
• Implementation
• Evaluation
• Communication
• Marketing
• Workplanning
So logic models can be used to help plan and
manage the whole program
Page 25
Logic Model Construction Process
Brainstorming the Draft Model:
• Establish the scope and context
• Determine challenge(s), outcomes, inputs,
activities, outputs, and measurement
components
Page 26
Logic Model Construction Process
Continued…
• Create model draft
• Express relationships among/between key
components
• Determine evaluation needs/points using dotted
line arrows (solid arrows show known
relationships)
Page 26 - A
Logic Model Construction Process
Brainstorm the Model Development:
•
•
•
•
Use flipcharts and colored post-its
Create model draft step by step
No idea is “wrong”
Think creatively!
Page 27
Logic Model Development
1. Identify CHALLENGE/Social ill: What do we
want to improve in the population?
• Out of wedlock births
• Relationships before marriage
• Diseases (STDs)
Challenges are often expressed as statements
of fact, based on empirical data/statistics.
There may be multiple challenges addressed
by a single program.
Page 28
Logic Model Development
2. Identify LONG-TERM OUTCOMES: The ultimate end
goals of your program, how your service population will
look after your interventions have taken place
• Decrease out of wedlock births
• Increase proportion of abstinent youths
• Decrease preventable disease (STDs)
Long-term outcomes/goals (as well as all other outcomes)
are usually expressed as changes: you will use words
such as “improved,” “increased,” “decreased,” etc.
Ultimate long-term outcomes/goals are sometimes “pie in
the sky” or utopian.
Page 29
Logic Model Development
The remainder of the logic model elaborates
how we get from the CHALLENGE to the
LONG TERM OUTCOMES.
CHALLENGE
INPUTS
ACTIVITIES
DEMOGRAPHICS
EXT. FACTORS
CONSTRAINTS
ASSUMPTIONS
OUTPUTS
OUTCOMES
LONG TERM
OUTCOMES
(GOALS)
Logic Model Development
3. Identify Inputs/Resources (personnel, funds,
laws/regulations, creative ideas, etc.)
4. Identify Activities:
• Train the trainer
• Delivering abstinence education in [churches,
community centers, schools]
Page 31
Logic Model Development
5. Identify Outputs, such as number of trainers trained,
number of training courses developed and
administered.
6. Identify/develop Key Outcomes and Measures:
• Decrease the rate of births to unmarried teenage girls
ages 15-19 (35.4% in 2002) Target: 35% in 2003
•
Decrease the proportion of youth ages 15-19 who
have engaged in sexual intercourse (46.7% in 2003)
Target: 45.5% for 2004
… and set Targets
Page 32
Logic Model Development
7. Determine where program evaluation/research
needs to take place (and depict arrows
accordingly):
- - - - - - > = plausible causal relationship
(or desired effects/results)
---------- = known causal relationship
(based on scientific research/program
evaluation)
Page 33
Logic Model Development
Reminders:
•
•
•
•
Logic modeling is a continuous (not static) process
Incorporate activities/processes into more detailed
project workplans
Do not restrict your thinking - jot down ideas for the
workplan or other areas, items, etc. as you think of
them while brainstorming
Use the abstinence-until-marriage (Abstinence
Education) logic model template as a starting
point, and to assist thinking...
Page 34
Logic Model Development
Complementary Tools:
• Workplans
• Strategic plans
• Flowcharts
• Process diagrams
• Related logic models
• Performance budgets containing “global”
Agency performance measure information
Page 35
LINKING – How does all this relate?
• Performance measurement & reporting (represented in the
boxes of the logic model): focus is on outcomes/ultimate
results
• Program evaluation (represented in the arrows of the logic
model): focus is on impacts
• Logic Modeling (gives a clear picture of what outcome
measures needed, program evaluation needed, etc.)
Page 36
In Conclusion
• Need to focus on producing results
• We have lots of information! – how do we
assemble it to help us manage?
• Use logic modeling to track and improve
performance (at federal, state, and
grantee levels)
Page 37
Questions / Discussion
• Questions on any aspect of
measuring program results?
• Is it clear how all aspects relate to
one another and how logic modeling
can be used as a powerful tool?
Page 38
Go Forth, Produce Results, and Measure !
. . . for our youth
Page 39
Questions & Answers
• You may submit questions pertaining to
today’s web cast until 5:00pm EDT,
Wednesday June 14, 2006, to the following
address:
cbaeta@youthdevelopment.org
• Answers will be posted at www.paltech.com/web/cbaewebcast/ as soon as
they are available.
Page 40
Download