Planning How to Conduct the Evaluation

advertisement
Planning How to Conduct
the Evaluation
Dr. Suzan Ayers
Western Michigan University
(courtesy of Dr. Mary Schutten)
Relationship of focus and planning
• The structure of evaluation consists of:
– focusing the evaluation
– collecting, organizing, analyzing & reporting info
– administering the evaluation (see CIPP notes)
• Evaluation should be conducted flexibly
• Evaluator needs clear understanding of evaluation’s
purpose and role
• Focusing the evaluation determines what information is
needed
– Origin and context of the proposed evaluation
– Identifying/selecting questions, criteria, standards
• Once the evaluation questions are known, the next
step is to determine what information is needed to
answer each question
• Planning the evaluation consists of collecting info,
organizing info, analyzing info, reporting info, and
administering the evaluation
SAMPLE
• Evaluation Question:Have the critical program
activities occurred on time and within budget?
• Things evaluator would need to know:
– Which activities were viewed as critical?
– Program time frames budget
– Program time frames budget by activity
– When each activity began/ended
– Total cost of each critical activity
Identifying Design and
Data Collection Methods
• Involve client and stakeholders in deciding necessary info to
best answer each question
• Designs specify organization/structure for data collection
• Causal designs: (quasi)experimental designs
– Descriptive case studies, multiple regression or other stats methods to answer
evaluation Qs
• Descriptive designs: describe (case study), analyze the program,
show a trend (time series), assess public opinions (cross-sectional),
illustrate a process (thick description)
– Commonly used in needs assessment and process studies
• Evaluator and stakeholders examine each question
carefully to identify any important research design
issues
• Most evaluations are multiple research designs or
combinations
• Important to discuss early to see if:
– groups are available, appropriateness of random
assignment, time collecting data, etc…
– Is the design “do-able?”
Identifying Appropriate
Information Sources
• Once needed information is agreed upon, the
source(s) of that information must be specified
– “Source”: group of individuals or location of
existing information that answers each Q
– Who will have information or access to it?
– Who will be able to collect those data?
• Using existing data as information source
– Does necessary information already exist in a
readily available form?
• Commonly used information sources
– Program recipients, deliverers, persons who have
knowledge of the program recipients, public
documents/databases
• Policies that restrict information sources
– Do policies exist concerning collecting data from
clients or existing files?
• Confidentiality, anonymity, privacy, IRB protocols
• Client involvement in identifying sources
– Evaluator, by training and experience, often can identify
key sources of information
– Client will be able to identify sources of information that
may be missed by the evaluator
– Collaboration yields helpful answers and enhances sense of
shared ownership
Identifying Data Collection Methods and
Instruments
• Data collected directly from individuals identified as
sources of information
– Self reports
• interviews, surveys, rating scales, focus groups,
logs/journals
– Personal Products:
• tests (objective, essay), performances (simulations, role-play, competency
testing), work samples (portfolios)
• Data collected by independent observer
– Narrative accounts
– Observation forms (rating scales, checklists)
• Data collected with technological device
– Audiotape
– Videotape
– Time-lapse photos
– Others
• BP cuffs, speed, graphic recordings of performance
skills, computer collation of participant responses
• Data collection from unobtrusive measures
• Data collected from existing information
– Public documents
• federal, state, local, databases etc..
– Review of organizational documents
• client files, notes of employees/directors, audits,
minutes, publications
– personal files
• Correspondence, e-mails, etc.
• After identifying for use some of the methods on the
next slides, it is important to review adequacy of
techniques
– Will the info collected provide a comprehensive
picture?
– Are the methods legal and ethical?
– Is cost of data collection be worthwhile?
– Can they be collected w/o undo disruption?
– Can data be collected w/in time constraints?
– Will the information be reliable and valid for the
purposes of the evaluation?
Determining Appropriate Conditions for
Collecting Information
• 3 key issues around data collection:
– Will sampling be used?
– How will data actually be collected?
– When will data be collected?
• Specifying sampling procedures to be employed
– Sampling helps researcher draw inferences about the
population in the study
– Sampling is a tool to use whenever resources and time are
limited
– Sampling useful when it will not diminish the confidence
of results
– Sample size must be appropriate; too small a sample is of
limited value
• Specifying how the information will be collected
– Who will collect data?
– For interviews, focus groups etc… how will characteristics
of the evaluator influence data collection?
– What training should be given to people collecting the
data?
– In what setting should data collection take place?
– Confidentiality protected?
– Special equipment, materials needed?
– Evaluators need a bigger ‘bag of tools’ needed than most
researchers to examine a broad array of phenomena
• Specifying when the information will be collected
– When will the information be needed?
– When will the information be available?
– When can the information conveniently be
collected?
Determining Appropriate Methods to
Organize, Analyze, Interpret Information
• Develop a system to code, organize, store, and retrieve data
• For each evaluation question, specify how collected
information will be analyzed
– Identify statistical/summarizing techniques
– Designate some means for conducting the analysis
• Interpreting results (Statistical reports do not speak for the themselves)
– Share information with clients to gain perspective on potential
interpretations of the data
– Criteria/Standards guide interpretation of some Qs
– Eval plan should allow recording of multiple or conflicting
interpretations
– Interpretations should consider multiple perspectives
Determining Appropriate Ways to
Report Evaluation Findings
• Using a matrix appropriate way to plan reporting (Fig 13.1)
– Audience, content, format, date, context of presentation
• Suggested Qs (Brknerhoff, Brethower, Hluchyj, & Nowakowski, 1983)
– Are reporting audiences defined?
– Are report formats and content appropriate for audience
needs?
– Will the evaluation report balanced information?
– Will reports be timely and efficient?
– Is the report plan responsive to the rights to info of the
audiences?
Appropriate Ways to Report
Evaluation Findings
Worksheet approach; summarize: 8 topics for each Q
1. Information required to answer the question
2. Design(s) to be used to collect information
3. Sources(s) of that information
4. Method(s) for information collection
5. Information-collecting arrangements
-sampling procedure
-collection procedure
-schedule for collection
6. Analysis procedures
7. Interpretation procedures (including standards)
8. Reporting procedures
-audiences
-content
-format
-schedule
-context
The Management Plan
• Final task in planning the evaluation study is
describing how it will be carried out
• Management plan essential to help oversee the project
– Who will do it?
– How much will it cost?
– Will it be within budget?
• Evaluation management is multifaceted
– supervise staff
– serve as liaison to evaluation clients, participants,
and stakeholders
– identify and cope with political influences
– communicate, communicate, communicate
• Evaluation, whether by one person or a team, cannot
afford to be disorganized or haphazard
• Management plan needed to structure and control
resources
• Good management plans specify for ea eval Q (Fig 13.4) :
– Tasks & timelines, personnel/resources for each task, cost
• Managing and estimating time for conducting
evaluation
– PERT (program eval & review technique) & Gantt charts are
commonly used to estimate time on tasks
– Gantt charts (Fig 13.6): simple displays that include
chronologically scaled time frames for each evaluation
task
• Y (vertical) axis: tasks
• X (horizontal) axis: time scale
• Horizontal line drawn for each task show time needed
– Help highlight interim deadlines or milestones that
must be met to stay on the time in the study
– Well-specified milestones essential monitoring tool
• Analyzing personnel needs and assignments
– Quality of evaluation depends heavily on those who carry
it out
– Are qualified individuals available to carry out the tasks?
• “Personnel role specifications” for all tasks (Suarez, 1981)
• Specify who would manage the study, complete the eval design,
select or develop instruments, collect data, analyze data, write
summary reports, etc…
• Estimating costs and developing budgets
– Staff salary and benefits
– Consultants
– Travel and per diem ( for staff/consultants)
– Communications (postage, phone, etc.)
– Printing and duplication
– Data processing
– Printed materials
– Supplies and equipment
– Subcontracts
– Overhead (facilities, utilities)
If initial estimate exceeds expectations, review each line item
• Agreements and contracts
• Potential problems that arise during the evaluation can
be more easily resolved if client and evaluator share a
firm understanding
• A well-documented agreement prior to launching the
evaluation study concerning important procedures is
very helpful
• http://www.wmich.edu/evalctr/checklists/contracts.pdf
Download