Planning your evaluation

advertisement
Planning
your
evaluation
This presentation provides an overview of the steps to
planning a quality evaluation. Each step is covered in
greater detail in other sections on the web site.
© 2009 University of Wisconsin-Extension, Cooperative Extension, Program Development and Evaluation
1
An evaluation plan is your roadmap
How do you know which way to go
if you don’t know where you are going?
© 2009 University of Wisconsin-Extension, Cooperative Extension, Program Development and Evaluation
2
Use the booklet,
Planning a Program
Evaluation and the
Worksheet as
background for this set
of slides and to help
with your evaluation
plan.
Links to these
resources are found on
the web page:
http://www.uwex.edu/ces/pdande/evaluation
© 2009 University of Wisconsin-Extension, Cooperative Extension, Program Development and Evaluation
3
There are
5 core
steps
in
planning a
program
evaluation
© 2009 University of Wisconsin-Extension, Cooperative Extension, Program Development and Evaluation
4
These are best practices steps!
Consider these steps regardless if you are
evaluating a single workshop or a
comprehensive program.
However, the level of detail in your evaluation
plan will depend upon the scope and nature of
your evaluation.
© 2009 University of Wisconsin-Extension, Cooperative Extension, Program Development and Evaluation
5
STEP 1
Begin by deciding who should
be involved in helping design and
implement the evaluation.
Or, do you do it alone?
© 2009 University of Wisconsin-Extension, Cooperative Extension, Program Development and Evaluation
6
Think about these 4-H YD stakeholders
as people who might be involved in your
evaluation ( your evaluation stakeholders)
•
•
•
•
•
•
Youth participants
Your program partners
Parents
Teachers
Volunteers
Your funder
© 2009 University of Wisconsin-Extension, Cooperative Extension, Program Development and Evaluation
7
How might they be involved?
What roles might they play?
They might
• Help determine what the evaluation will
focus on –
what you really need to know
• Be part of an evaluation team
• Help write the questions
• Collect data
• Help enter data and/or do the analyses
• Raise funds
• Write a press releases
© 2009 University of Wisconsin-Extension, Cooperative Extension, Program Development and Evaluation
8
Why would you want to do this?
• To build ownership of the evaluation and your
work
• To develop skills in others, e.g., youth who
learn how to conduct a survey or analyze data
• To ensure that the evaluation findings will be
useful and used
• To bring talent and expertise to the evaluation
• To share the work!
© 2009 University of Wisconsin-Extension, Cooperative Extension, Program Development and Evaluation
9
STEP 2
Next, we focus the evaluation.
What program or part of a program are you
going to evaluate?
Use a logic model to describe your program so
you evaluate what is meaningful.
© 2009 University of Wisconsin-Extension, Cooperative Extension, Program Development and Evaluation
10
A logic model helps us clearly describe the
program we want to evaluate. It helps us focus
on what we want to collect information about.
Situation:
INPUTS
Program
investments
OUTPUTS
Activities
Assumptions
Participation
OUTCOMES
Short
Medium
Long-term
External factors
© 2009 University of Wisconsin-Extension, Cooperative Extension, Program Development and Evaluation
11
Be clear about your
purpose for evaluating
We are conducting an evaluation of
______________ (name of program)
because _________________________
in order to ________________________.
Example:
We are conducting an evaluation of the 3 series
Money Quest Program because we want to know to
what extent youth participants increase their
knowledge about money management and use the
recommended practices in order to report the value
of this program to our funder.
© 2009 University of Wisconsin-Extension, Cooperative Extension, Program Development and Evaluation
12
Determine Use and User
WHO?
WHAT
do you/they
want to know?
HOW
will they use the
info?
You – staff
Participants
Funder
Other
stakeholders ??
© 2009 University of Wisconsin-Extension, Cooperative Extension, Program Development and Evaluation
13
What do you want to know –
what data do you want to collect?
Go back to your logic model – the description of your
program and think about what you really want to know
about your program. What data do you want to collect?
Do you want information about:
• Outcomes: to what extent changes occur in
knowledge, skills, attitudes, opinions, behaviors,
practices, policies, social-economic-environmental
quality…
• Reactions: What the participants like/dislike?
Whether they will come again/stay involved/promote
your program to others, etc. ?
© 2009 University of Wisconsin-Extension, Cooperative Extension, Program Development and Evaluation
14
Do you want to know about …
• Participation
Who/how many attended? Who didn’t attend why? What happens for different participants?
• Activities
What was done and how well was it done (quality)
Did everything go according to plan? What
worked well/not so well? Which activities link to
different outcomes for different participants?
• Inputs
What was invested (inputs)? By whom? Who
collaborated? How? What resources were used?
© 2009 University of Wisconsin-Extension, Cooperative Extension, Program Development and Evaluation
15
Do you want to know about…
• Your teaching?
How you did? How you might improve?
• Future programming?
Future interests; needs?
• Costs and returns?
How much it costs to put the program on and
what the return on that investment is?
• Other questions???
© 2009 University of Wisconsin-Extension, Cooperative Extension, Program Development and Evaluation
16
Prioritize:
We can’t and don’t want to collect information
about everything!
© 2009 University of Wisconsin-Extension, Cooperative Extension, Program Development and Evaluation
17
How will you know it? INDICATORS
For each evaluation question, think about the
specific information you need to collect to answer
that question.
Example:
Question
Indicator
To what extent did the
program increase youthadult partnerships?
- #,% of Boards with youth participating
in meetings before and after
- #,% of Boards with youth on
committees before and after
- #,% of Boards with youth in
leadership positions before and after
- Reported change in quality of the
youth-adult interaction
© 2009 University of Wisconsin-Extension, Cooperative Extension, Program Development and Evaluation
18
What is your “evaluation design”
Evaluation design is your overall approach to
collecting data.
Typical evaluation designs include :
– Single point in time (e.g., survey, end-of-program
questionnaire)
– Pre-post program or retrospective post-then-pre
(comparison of before to after)
– Multiple points in time (e.g., pre, post and follow-up
– Comparison group designs (two or more groups)
– Case study design (indepth analysis of a single
case)
– Mixed method (e.g., the use of a survey, observation
and testimonials)
© 2009 University of Wisconsin-Extension, Cooperative Extension, Program Development and Evaluation
19
STEP 3
Now, think about from whom and what method
you will use to collect the information.
© 2009 University of Wisconsin-Extension, Cooperative Extension, Program Development and Evaluation
20
From where or from whom will you get the
information you need? (“Source”)
Do you think that you are most likely to get the
information you need from:
– Existing information – records, reports,
program documents, etc.
– People – participants, parents, volunteers,
etc.
– Pictorial records and observations – video
or photos, observations of events
© 2009 University of Wisconsin-Extension, Cooperative Extension, Program Development and Evaluation
21
What method(s) of data collection is most
appropriate ?
• Survey
(questionnaire)
• Case study
• Interview
• Observation
• Group assessment
• Expert or peer
reviews
• Portfolio reviews
• Testimonials
• Tests
• Photographs,
videotapes, slides
• Diaries, journals, logs
• Document review and
analysis
© 2009 University of Wisconsin-Extension, Cooperative Extension, Program Development and Evaluation
22
Which method do you choose?
As we know, there are many standard methods
of data collection, and more creative ones. Each
has its own strengths and weaknesses.
The ‘art and science’ of data collection is to
select the method appropriate for the purpose of
your evaluation, the audience you are collecting
information from and your resources.
© 2009 University of Wisconsin-Extension, Cooperative Extension, Program Development and Evaluation
23
Test your understanding:
Match the method in the left column to
the definitions on the right
• Case study
• Questionnaire
• Observation
• Focus group
• Interview
1. Collecting standardized
information from people in a
non-threatening way
2. Talking with and listening to
people either face-to-face or by
telephone.
3. Gathering information by
viewing what is occurring
4. Exploring a topic in depth
through group discussion
5. Gaining an in-depth
understanding of someone’s
experience in the program
© 2009 University of Wisconsin-Extension, Cooperative Extension, Program Development and Evaluation
24
STEP 4
While the actual analysis comes once you have
your data, you want to think about analysis when
you are planning the evaluation. That way you will
be sure to collect the type of information to be able
to report as you want.
© 2009 University of Wisconsin-Extension, Cooperative Extension, Program Development and Evaluation
25
What do you want to be able to report?
Will you want to report
Type of analysis
Numbers, percents
Count; percentage
Average number or score
Range of scores
Mean
Range
Changes from before to after
Change score
Comparisons of one group to
another
Cross tab
People’s stories
Qualitative content analysis
Comments
Qualitative content analysis
© 2009 University of Wisconsin-Extension, Cooperative Extension, Program Development and Evaluation
26
Step 5
And, last but not least, is USING and
COMMUNICATING your evaluation findings. Actually,
we should be communicating and using our
information throughout the evaluation…not just at the
end.
Think about the many ways you might use and share
information both during and after your evaluation.
© 2009 University of Wisconsin-Extension, Cooperative Extension, Program Development and Evaluation
27
How will you report, use, and learn from
the findings?
WHO - To whom will you report?
WHAT - What will you report/share?
HOW - How will you share the information?
WHEN - When will you communicate?
© 2009 University of Wisconsin-Extension, Cooperative Extension, Program Development and Evaluation
28
Managing the evaluation
We might think about a management
plan to make sure everything gets done!
– Human subjects protection
Check the decision tree and follow best practice
guidelines. Consult with the Human Subjects
Administrator if you have questions:
http://www.uwex.edu/hsp/
– Timeline
– Responsibilities: who will do what?
– Budget: money needed to do the
evaluation
© 2009 University of Wisconsin-Extension, Cooperative Extension, Program Development and Evaluation
29
Check your evaluation plan against the
Evaluation Standards to ensure a ‘quality’
evaluation
• Utility: Will it be useful and used?
• Feasibility: Is it practical and can be
accomplished given your resources?
• Propriety: Is it respectful and ethical?
• Accuracy: Is it likely to produce accurate
information?
© 2009 University of Wisconsin-Extension, Cooperative Extension, Program Development and Evaluation
30
Remember…
“The more attention you give to
planning the evaluation,
the more effective it will be.”
−The Program Manager’s Guide to Evaluation, 2003
© 2009 University of Wisconsin-Extension, Cooperative Extension, Program Development and Evaluation
31
Reflection time
What is one thing
that you learned from
this presentation that
you didn’t know before?
Good luck
with planning your next evaluation!
© 2009 University of Wisconsin-Extension, Cooperative Extension, Program Development and Evaluation
32
Download