Building Evaluation Capacity Presentation Slides for Participatory Evaluation Essentials:A Guide for Non-Profit Organizations And Their Evaluation Partners Anita M. Baker, Ed.D. Bruner Foundation Rochester, New York How to Use the Bruner Foundation Guide & Powerpoint Slides Evaluation Essentials:A Guide for Nonprofit Organizations and Their Evaluation Partners. (the Guide) and slides are organized to help an evaluation trainee walk through the process of designing an evaluation and collecting and analyzing evaluation data. The Guide also provides information about writing an evaluation report. The slides allow for easy presentation of the content, and in each section of the Guide there are activities that provide practice opportunities. The Guide has a detailed table of contents for each section and it includes an evaluation bibliography. Also included are comprehensive appendices which can be pulled out and used for easy references, as well as to review brief presentations of other special topics that are not covered in the main section and sample logic models, completed interviews which can be used for training activities, and a sample observation protocol. For the Bruner Foundation-sponsored REP project, we worked through all the information up front, in a series of comprehensive training sessions. Each session included a short presentation of information, hands-on activities about the session topic, opportunities for discussion and questions, and homework for trainees to try on their own. By the end of the training sessions, trainees had developed their own evaluation designs which they later implemented as part of REP. We then provided an additional 10 months of evaluation coaching and review while trainees actually conducted the evaluations they had designed and we worked through several of the additional training topics that are presented in the appendix. At the end of their REP experience, trainees from non-profit organizations summarized and presented the findings from the evaluations they had designed and conducted. The REP non-profit partners agreed that the up-front training helped prepare them to do solid evaluation work and it provided opportunities for them to increase participation in evaluation within their organizations. The slides were first used in 2006-07 in a similar training project sponsored by the Hartford Foundation for Public Giving. We recommend the comprehensive approach for those who are interested in building evaluation capacity. Whether you are a trainee or a trainer, using the guide to fully prepare for and conduct evaluation or just look up specific information about evaluation-related topics, we hope that the materials provided here will support your efforts. Bruner Foundation Rochester, New York These materials are for the benefit of any 501c3 organization. They MAY be used in whole or in part provided that credit is given to the Bruner Foundation. They may NOT be sold or redistributed in whole or part for a profit. Copyright © by the Bruner Foundation 2007 * Please see the notes attached to the first slide for further information about how to use the available materials. Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. Building Evaluation Capacity Session 1 Important Definitions Thinking About Evaluative Thinking Anita M. Baker, Ed.D. Bruner Foundation Rochester, New York Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 4 Working Definition of Program Evaluation The practice of evaluation involves thoughtful, systematic collection and analysis of information about the activities, characteristics, and outcomes of programs, for use by specific people, to reduce uncertainties, improve effectiveness, and make decisions. Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 5 Working Definition of Program Evaluation The practice of evaluation involves thoughtful, systematic collection and analysis of information Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 6 Working Definition of Program Evaluation The practice of evaluation involves thoughtful, systematic collection and analysis of information about the activities, characteristics, and outcomes of programs, Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 7 Working Definition of Program Evaluation The practice of evaluation involves thoughtful, systematic collection and analysis of information about the activities, characteristics, and outcomes of programs, for use by specific people, Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 8 Working Definition of Program Evaluation The practice of evaluation involves thoughtful, systematic collection and analysis of information about the activities, characteristics, and outcomes of programs, for use by specific people, to reduce uncertainties, improve effectiveness, and make decisions. Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 9 Working Definition of Participatory Evaluation Participatory evaluation involves trained evaluation personnel and practice-based decision-makers working in partnership. It brings together seasoned evaluators with seasoned program staff to: Address training needs Design, conduct and use results of program evaluation Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 10 Evaluation Strategy Clarification All Evaluations Are: Partly social Partly political Partly technical Both qualitative and quantitative data can be collected and used and both are valuable There are multiple ways to address most evaluation needs. Different evaluation needs call for different designs, types of data and data collection strategies. Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 11 Purposes of Evaluation Evaluations are conducted to: Render judgment Facilitate improvements Generate knowledge Evaluation purpose must be specified at the earliest stages of evaluation planning and with input from multiple stakeholders. Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 12 What is an Evaluation Design? An Evaluation Design communicates plans to evaluators, program officials and other stakeholders. Evaluation Designs help evaluators think about and structure evaluations. Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 13 Good Evaluation Designs Include the Following Summary Information about the program The questions to be addressed by the evaluation The data collection strategies that will be used The individuals who will undertake the activities When the activities will be conducted The products of the evaluation (who will receive them and how they should be used) Projected costs to do the evaluation Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 14 Evaluation Questions Get you Started Focus and drive the evaluation. Should be carefully specified and agreed upon in advance of other evaluation work. Generally represent a critical subset of information that is desired. Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 15 What about Evaluation Stakeholders? Evaluation stakeholders include anyone who makes decisions about a program, desires information about a program, and/or is involved directly with a program. • Most programs have multiple stakeholders. • Stakeholders have diverse, often competing interests. Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 16 Who are Evaluation Stakeholders? Organization officials Program staff Program clients or their caregivers Program Funders Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 17 What do you need to know about a program …. before you design an evaluation? 1. What is/are the purpose(s) of the program? 2. What stage is the program in? 3. Who are the program clients? 4. Who are the key program staff (and where applicable, in which department is the program? 5. What specific strategies are used to deliver program services? 6. What outcomes are program participants expected to achieve? 7. Are there any other evaluation studies currently being conducted regarding this program? 8. Who are the funders of the program? 9. What is the total program budget? 10. Why has this program been selected for evaluation? Bruner Foundation Rochester, New York (new developing, mature, phasing out) Anita M. Baker, Ed.D. 18 Thinking About Evaluative Thinking Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 19 What is Evaluative Thinking? Evaluative Thinking is a type of reflective practice that incorporates use of systematically collected data to inform organizational decisions and other actions. Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 20 What Are Key Components of Evaluative Thinking? 1. Asking questions of substance 2. Determining data needed to address questions 3. Gathering appropriate data in systematic ways 4. Analyzing data and sharing results 5. Developing strategies to act on findings Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 21 What Are Key Components of Evaluative Thinking? 1. Asking questions of substance 2. Determining data needed to address questions 3. Gathering appropriate data in systematic ways 4. Analyzing data and sharing results 5. Developing strategies to act on findings Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 22 What Are Key Components of Evaluative Thinking? 1. Asking questions of substance 2. Determining data needed to address questions 3. Gathering appropriate data in systematic ways 4. Analyzing data and sharing results 5. Developing strategies to act on findings Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 23 What Are Key Components of Evaluative Thinking? 1. Asking questions of substance 2. Determining data needed to address questions 3. Gathering appropriate data in systematic ways 4. Analyzing data and sharing results 5. Developing strategies to act on findings Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 24 What Are Key Components of Evaluative Thinking? 1. Asking questions of substance 2. Determining data needed to address questions 3. Gathering appropriate data in systematic ways 4. Analyzing data and sharing results 5. Developing strategies to act on findings Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 25 What Are Key Components of Evaluative Thinking? 1. Asking questions of substance 2. Determining data needed to address questions 3. Gathering appropriate data in systematic ways 4. Analyzing data and sharing results 5. Developing strategies to act on findings Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 26 How is Evaluative Thinking Related to Organizational Effectiveness? Organizational capacity areas (i.e., core skills and capabilities, such as leadership, management, finance and fundraising, programs and evaluation) where evaluative thinking is less evident, are also capacity areas of organizations that usually need to be strengthened. Assessing evaluative thinking provides insight for organizational capacity enhancement. Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 27 Why Assess Evaluative Thinking? Assessment of Evaluative Thinking . . . helps clarify what evaluative thinking is helps to identify organizational capacity areas where evaluative thinking is more or less prominent (or even non-existent) informs the setting of priorities regarding how to enhance or sustain evaluative thinking Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 28 What Organizational Capacity Areas Does the Bruner Foundation Evaluative Thinking Tool Address? Mission Strategic Planning Governance Finance Leadership Fund Development Evaluation Client Relationships Bruner Foundation Rochester, New York Program Development Communication & Marketing Technology Acquisition & Training Staff Development Human Resources Alliances/Collaborations Business Venture Dev. Anita M. Baker, Ed.D. 29 How Can Evaluative Thinking be Assessed? Develop or locate a tool. Decide on an administrative approach and strategy: – Individual vs. Team/Group – Timing of administration – Communicating about the assessment Discuss how results could be used and plan for next steps. Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 30 The Bruner Foundation Evaluative Thinking Assessment Tool ORGANIZATION MISSION Assessment a. The mission statement is specific enough to provide a basis for developing goals and objectives b. The mission is reviewed and revised on a scheduled basis (e.g. annually) with input from key stakeholders as appropriate c. The organization regularly assesses compatibility between programs and mission d. The organization acts on the findings of compatibility assessments (in other words, if a program is not compatible with the mission, it is changed or discontinued) Comments: Priority #DIV/0! Please proceed to the next Worksheet Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 31 The Bruner Foundation Evaluative Thinking Assessment Tool GOVERNANCE Assessment a. Board goals/workplan/structure are based on the mission and strategic planning b. Board uses evaluation data in defining goals/workplan/structure and organizational strategic planning c. Board regularly evaluates progress relative to own goals/workplan/structure d. There is a systematic process and timeline for identifying, recruiting, and electing new board members e. Specific expertise needs are identified and used to guide board member recruitment f. The board regularly (e.g., annually) evaluates the executive director’s performance based on established goals/workplan g. Board members assess and approve the personnel manual covering personnel policy h. The board assess the organization’s progress relative to long-term financial plans i. The board assess the organization’s progress relative to program evaluation results Comments: Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. Priority #DIV/0! 32 The Bruner Foundation Evaluative Thinking Assessment Tool TECHNOLOGY ACQUISITION PLANNING AND TRAINING Assessment a. An assessment process is in place to make decisions about technology maintenance, upgrades, and acquisition b. Technology systems include software that can be used to manage and analyze evaluation data (e.g., Excel, SPSS) c. Technology systems provide data to evaluate client outcomes d. Technology systems provide data to evaluate organizational management e. Technology systems are regularly assessed to see if they support evaluation f. Staff technology needs are regularly assessed Comments: Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. Priority #DIV/0! 33 The Bruner Foundation Evaluative Thinking Assessment Tool Bruner Foundation Evaluative Thinking Assessment Organizational Capacity Area Capacity Score* Action Planning** (Select from list) 1 Mission 50 Action suggested see priorities 2 Strategic Planning 50 Action suggested see priorities 3 Governance 63 No action required in this area 4 Leadership 92 No action required in this area 5 Finance 71 Action suggested see priorities 6 Fund Development/Fund Raising 50 Action suggested see priorities 7 Evaluation 69 Action required see priorities 8 Program Development 80 No action required in this area 9 Client Relationships 80 No action required in this area 10 Communication and Marketing 80 No action required in this area 11 Technology Acquisition and Planning 67 Action suggested see priorities 12 Staff Development 67 Action suggested see priorities 13 Human Resources 33 Action required see priorities 14 Business Venture Development 50 No action required in this area 15 Alliances and Collaboration 40 No action required in this area Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 34 The Bruner Foundation Evaluative Thinking Assessment Tool Evaluative Thinking Scores 40 Alliances and Collaboration 50 Business Venture Development 33 Human Resources 67 Staff Development 67 Technology Acquisition and Planning Communication and Marketing 80 Client Relationships 80 Program Development 80 69 Evaluation 50 Fund Development/Fund Raising 71 Finance 92 Leadership 63 Governance Strategic Planning 50 Mission 50 0 Bruner Foundation Rochester, New York 10 20 30 Anita M. Baker, Ed.D. 40 50 60 70 80 90 100 35 How Should Evaluative Thinking Assessment Results be Used? 1. Review assessment results. 2. Distinguish communications vs. strategic issues (where possible). 3. Identify priorities and learn more about strategies to enhance Evaluative Thinking. 4. Develop an action plan based on priorities and what’s been learned about enhancing Evaluative Thinking. 5. Re-assess Evaluative Thinking and determine the effectiveness of the action plan. Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 36 Building Evaluation Capacity Session 2 Logic Models Outcomes, Indicators and Targets Anita M. Baker, Ed.D. Bruner Foundation Rochester, New York Logic Model Overview Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 38 So, what is a logic model anyway? A Logic Model is a simple description of how a program is understood to work to achieve outcomes for participants. It is a process that helps you to identify your vision, the rationale behind your program, and how your program will work. Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 39 And . . . . Logic models are useful tools for program planning, evaluation and fund development. Developing or summarizing a logic model is a good way to bring together a variety of people involved in program planning to build consensus on the program’s design and operations. Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 40 Why use a Logic Model? Developing a Logic Model will help you get clear about what you’re doing, and how you hope it will make a difference. You have the best knowledge of the context of your work and what’s important to you and your communities. Developing a Logic Model draws from what you already know. A Logic Model will leave you with a clear, thoughtful plan for what you are doing and what you hope to achieve. This plan can be an advocacy resource, bring clarity to your message and help you tell your story. Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 41 To Construct a Logic Model You Must Describe: Inputs: resources, money, staff/time, facilities, etc. Activities: how a program uses inputs to fulfill its mission – the specific strategies, service delivery. Outcomes: changes to individuals or populations during or after participation. Inputs Bruner Foundation Rochester, New York Activities Anita M. Baker, Ed.D. Outcomes 42 Here is an illustration that will help you create your own Logic Model. Contextual Analysis Identify the major conditions and reasons for why you are doing the work in your community Inputs Resources dedicated to or consumed by the program. E.G. money staff and staff time, volunteers and volunteer time facilities equipment and supplies Activities What the program does with the inputs to fulfill its mission. Outcomes Benefits for participants during and after program activities. E.G. provide x number of classes to x participants provide weekly counseling sessions educate the public about signs of child abuse by distributing educational materials to all agencies that serve families Identify 20 mentors to work with youth and opportunities for them to meet monthly for one year E.G. new knowledge increased skills changed attitudes modified behavior improved condition altered status Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 43 Let’s analyze an example logic model Contextual Analysis • Jobs exist, we just have to help people find them. The absence of a job history perpetuates unemployment. •Education can help people improve their skills. Being able to ask a mentor for advice is useful. •Job seekers need help with soft skills and technical training. • Personal, one-on-one attention and classes can inspire and support people in keeping jobs and establishing job histories. Short-term Outcomes Activities • Provide 6 weekly Soft Skills classes. •Identify on-the-job training opportunities and assist participants with placement. • Conduct 6 months of on-the-job supervised training and lunchtime mentoring sessions Bruner Foundation Rochester, New York People in my community: • Have few job skills and are likely to have bad jobs or no jobs, and limited job histories. • Have few opportunities for job training, placement, or help to deal with issues that come up while on the job. -Participants learn Assumptions specific marketable skills and strategies Getting solid hard and soft to help them get and skills are the keep jobs. first steps to -Participants establish keeping a job. trusting relationships If people feel with mentors who can supported, they will keep answer questions and working. support them while they are involved in on-the-job training. Anita M. Baker, Ed.D. Ask yourself…. …do the outcomes seem reasonable given the program activities? …do the assumptions resonate with me and my experiences? …are there gaps in the strategy? Longer-term Outcomes - Participants maintain their employment and establish records that increase the likelihood for continuous work and better jobs. 44 Summarizing a Logic Model Helps to: Create a snapshot of program operations that shows what is needed, how services are delivered and what is expected for participants. Describe programs currently or optimally. Identify key components to track. Think through the steps of participant progress and develop a realistic picture of what can be accomplished. Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 45 Important Things to Remember Not all programs lend themselves easily to summarization in this format. Logic models are best used in conjunction with other descriptive information or as part of a conversation. It is advisable to have one or two key project officials summarize the logic model but then to have multiple stakeholders review it and agree upon what is included and how. Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 46 Important Things to Remember When used for program planning, it is advisable to start with outcomes and then determine what activities will be appropriate and what inputs are needed. There are several different approaches and formats for logic models. This one is one-dimensional and limited to three program features (inputs, activities, outcomes). The relationships between inputs, activities and outcomes are not one-to-one. The model is supposed to illustrate how the set of inputs could support the set of activities that contribute to the set of outcomes identified. (Levels of service delivery or “outputs” are shown in the activities.) Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 47 Use Logic Models for Planning, Evaluation and Fund Development Contextual Analysis What is needed to address the context that exists? Identify the major conditions and reasons for why you are What would be interesting to try? doing or could do this work What do we need to respond to this RFP? Ask yourself…. …do the outcomes seem reasonable given the program activities? …do the assumptions resonate with me and my experiences? …are there gaps in the strategy? Short-term Inputs: What resources do we need, can we dedicate, or do we currently use for this project? Activities: What can or do we do with these inputs to fufill the program mission? Bruner Foundation Rochester, New York Outcomes: Assumptions What benefits for participants during and When do we after the program can think outcomes we or do we expect? New knowledge? Increased skills? Changed attitudes? Modified behavior? Improved condition? Altered status? Anita M. Baker, Ed.D. will happen – will what happens initially affect or cause other longer-term outcomes? How does this fit into our outcome desires overall? Longer-term Outcomes: What do we think happens ultimately? How does or can this contribute to organizational and community value? 48 Outcomes, Indicators and Targets Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 49 What is the difference between outcomes, indicators, and targets? Outcomes are changes in behavior, skills, knowledge, attitudes, condition or status. Outcomes are related to the core business of the program, are realistic and attainable, within the program’s sphere of influence, and appropriate. Outcomes are what a program is held accountable for. Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 50 What is the difference between outcomes, indicators, and targets? Indicators are specific characteristics or changes that represent achievement of an outcome. Indicators are directly related to the outcome and help define it. Indicators are measurable, observable, can be seen, heard or read, and make sense in relation to the outcome whose achievement they signal. Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 51 What is the difference between outcomes, indicators, and targets? Targets specify the amount or level of outcome attainment that is expected, hoped for or required. Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 52 Why measure outcomes? To see if your program is really making a difference in the lives of your clients To confirm that your program is on the right track To be able to communicate to others what you’re doing and how it’s making a difference To get information that will help you improve your program Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 53 Use Caution When Identifying Outcomes There is No right number of outcomes. Be sure to think about when to expect outcomes. 1)Initial Outcomes First benefits/changes participants experience 2)Intermediate Outcomes Link initial outcomes to longer-term outcomes 3)Longer-term Outcomes Ultimate outcomes desired for program participants Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 54 Use Caution When Identifying Outcomes Outcomes should not go beyond the program’s purpose. Outcomes should not go beyond the scope of the target population. Avoid holding a program accountable for outcomes that are tracked and influenced largely by another system. Do not assume that all subpopulations will have similar outcomes. Consider carefully unintended and possibly negative outcomes. Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 55 Identifying Outcomes: Consider This . . . • Is it reasonable to believe the program can influence the outcome in a non-trivial way? • Would measurement of the outcome help identify program successes and help pinpoint and address problems or shortcomings? • Will the program’s various “publics”–accept this as a valid outcome of the program? • Do program activities and outcomes relate to each other logically? GET FEEDBACK Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 56 How do you identify indicators? Indicators are specific characteristics or changes that represent achievement of an outcome. Indicators are directly related to the outcome and help define it. Indicators are measurable, observable, can be seen, heard or read, and make sense in relation to the outcome whose achievement they signal. Ask the questions shown on the following slide. Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 57 Questions to Ask When Identifying Indicators 1. What does this outcome look like when it occurs? 2. What would tell us it has happened? 3. What could we count, measure or weigh? 4. Can you observe it? 5. Does it tell you whether the outcome has been achieved? Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 58 Let’s “break it down” Use the “I’ll know it when I see it” rule The BIG question is what evidence do we need to see to be convinced that things are changing or improving? The “I’ll know it (outcome) when I see it (indicator)” rule in action -- some examples: I’ll know that retention has increased among home health aides involved in a career ladder program when I see a reduction in the employee turnover rate among aides involved in the program and when I see Bruner Foundation Rochester, New York survey results that indicate that aides are experiencing increased job satisfaction Anita M. Baker, Ed.D. 59 “I’ll know it when I see it” I’ll know that economic stability has increased among the clients I place in permanent employment when I see an increase in the length of time that clients keep their jobs and when I see an increase in the number of clients who qualify for jobs with benefits I’ll know my clients are managing their nutrition and care more effectively when I see my clients consistently show up for scheduled medical appointments and when I see decreases in my clients’ Bruner Foundation Rochester, New York body mass indexes (BMI) Anita M. Baker, Ed.D. 60 Remember! When Identifying Indicators . . . Indicators must be observable and measurable Indicators may not capture all aspects of an outcome. Many outcomes have more than one indicator. Identify the set that you believe (or have agreed) adequately and accurately signals achievement of an outcome. Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 61 Examples of Indicators Outcomes Initial: Teens are knowledgeable of prenatal nutrition and health guidelines Indicators Program participants are able to identify food items that are good sources of major dietary requirements Participants are within proper ranges for prenatal weight gain Intermediate: Teens follow proper nutrition and health guidelines Participants abstain from smoking Participants take prenatal vitamins Intermediate: Teens deliver healthy babies Bruner Foundation Rochester, New York Newborns weigh at least 5.5 pounds and score 7 or above on the APGAR scale. Anita M. Baker, Ed.D. 62 What are Targets ? Targets specify the amount or level of outcome attainment that is expected, hoped for or required. Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 63 How do you Identify Targets? Targets or levels of outcome attainment can be determined relative to: External standards (when they are available) Internal agreement • best professional hunches • past performance • performance of similar programs Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 64 Example of a Target Outcome: Parents will read to their preschoolers more often. Indicator: Parent reports of increased reading time after coming to the program. Target: 75% of participating parents will report a 50 percent increase in how often they read to their preschoolers. Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 65 Example of a Target Outcome: Parents will read to their preschoolers more often. Indicator: Parent reports of increased reading time after coming to the program. Target: 75% of participating parents will report reading to their preschoolers for at least 15 minutes, 4 or more nights per week. Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 66 Targets Some Words of Caution Performance targets should be specified in advance (i.e., when deciding to measure outcomes). Be sure there is buy-in regarding what constitutes a positive outcome – when the program has achieved the target and when it has missed the mark. Lacking data on past performance it may be advisable to wait. Be especially cautious about wording numerical targets so they are not over or under ambitious. Be sure target statements are in sync with meaningful program time frames. Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 67 Building Evaluation Capacity Session 3 Evaluation Questions and Designs Documenting Service Delivery Enhancing Service Delivery Anita M. Baker, Ed.D. Bruner Foundation Rochester, New York To Construct a Logic Model You Must Describe: Inputs: resources, money, staff/time, facilities, etc. Activities: how a program uses inputs to fulfill its mission – the specific strategies, service delivery. Outcomes: changes to individuals or populations during or after participation. It’s easiest to embed targets here Indicators: Indicators are specific characteristics or changes that represent achievement of an outcome Targets: specify the amount or level of outcome attainment that is expected, hoped for or required. Inputs Activities Bruner Foundation Rochester, New York Outcomes Anita M. Baker, Ed.D. Indicators w/ Targets Data Sources 69 To Construct a Logic Model You Must Describe: Inputs: resources, money, staff/time, facilities, etc. Activities: how a program uses inputs to fulfill its mission – the specific strategies, service delivery. Outcomes: changes to individuals or populations during or after participation. It’s easiest to embed targets here Indicators: Indicators are specific characteristics or changes that represent achievement of an outcome Targets: specify the amount or level of outcome attainment that is expected, hoped for or required. Inputs Activities Outcomes Indicators Data Sources Reports – staff, clients, sig. others Existing records – staff, clients Observation – staff, clients Test Results – staff, clients Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 70 Evaluation Strategy Clarification All Evaluations Are: Partly social Partly political Partly technical Both qualitative and quantitative data can be collected and used and both are valuable There are multiple ways to address most evaluation needs. Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 71 What is an Evaluation Design? An Evaluation Design communicates plans to evaluators, program officials and other stakeholders. Evaluation Designs help evaluators think about and structure evaluations. Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 72 Good Evaluation Designs Include the Following (see Appendix 6) Summary Information about the program The questions to be addressed by the evaluation The data collection strategies that will be used The individuals who will undertake the activities When the activities will be conducted The products of the evaluation (who will receive them and how they should be used) Projected costs to do the evaluation Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 73 Evaluation Questions . . . Focus and drive the evaluation. Should be carefully specified and agreed upon in advance of other evaluation work. Generally represent a critical subset of information that is desired. Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 74 Evaluation Questions: Criteria • It is possible to obtain data to address the • • • • questions. There is more than one possible “answer” to the question. The information to address the questions is wanted and needed. It is known how resulting information will be used internally (and externally). The questions are aimed at changeable aspects of programmatic activity. Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 75 Evaluation Questions: Advice Limit the number of evaluation questions Between two and five is optimal Keep it Manageable Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 76 Evaluation Questions: Examples Eval. Questions Data Collection/Protocol Questions How was staff training delivered, how did participants respond and how have they used what they learned? 1. How would you rate the staff training you received? 2. Did the staff training you received this year meet your needs? 3. Has the training you received changed your practice? 4. Has the training you received lead to changes in . . . Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 77 Evaluation Questions: Examples Eval. Questions Data Collection/Protocol Questions How was staff training delivered, how did participants respond and how have they used what they learned? 1. How would you rate the staff training you received? 2. Did the staff training you received this year meet your needs? 3. Has the training you received changed your practice? 4. Has the training you received lead to changes in . . . How and to what extent has the program met its implementation goals? 1. What does the X program do best? . What is your greatest concern? 2. Do staff communicate with caretakers as often as required? 3. Did you receive all the services promised in the program brochure? 4. How knowledgeable are staff about the issues you face? What impact has the program had on participants? 1. Have you changed the way you proceed with planning requirements? 2. Do you know more about guardianship now than before the program. 3. How would you rate this program overall? Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 78 Switching Gears How are Evaluative Thinking and Service Delivery (Activities) related? Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 79 When Organizations use Evaluative Thinking . . . • Client interaction includes collection and use of information. • Service delivery and program development include collection and use of information. Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 80 Examples of Evaluative Thinking: Client Interaction • Client needs assessments are conducted regularly. • Program services reflect client needs. • Client satisfaction and program outcomes are regularly assessed. • Results of client outcome assessments and client satisfaction are used. Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 81 Evaluative Thinking Client Data 1) Designing programs based on what funders want or only on what is “thought” to be best. 2) Assuming clients happiness = program effectiveness. 3) Collecting but not analyzing client data. 4) Limiting data collection from clients to satisfaction only. Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 82 Examples of Evaluative Thinking: Program Development • Identifying gaps in community services before planning new programs. • Assessing the needs of the target population as part of program planning process. • Using data from needs assessments and/or gaps analyses to inform planning. Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 83 Organizations that Regularly use Evaluative Thinking Will Also . . . • Think carefully about developing and assessing programs. • Incorporate program evaluation findings into the program planning. • Involve significant others in planning/revising. • Develop written program plans and logic models. • Follow program plans. • Have strategies in place to modify plans Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 84 What Can Organizational Leaders do to Enhance Evaluative Thinking? • Educate staff about Evaluative Thinking. • Be clear about what it means to take an evaluative approach. • Set the stage for others by using Evaluative Thinking in your own practice. Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 85 Remember Logic Models??? Inputs: resources, money, staff/time, facilities, etc. Activities: how a program uses inputs to fulfill its mission – the specific strategies, service delivery. Outcomes: changes to individuals or populations during or after participation. It’s easiest to embed targets here Indicators: Indicators are specific characteristics or changes that represent achievement of an outcome Inputs Bruner Foundation Rochester, New York Activities Outcomes Anita M. Baker, Ed.D. Indicators 86 What strategies are used to collect data about indicators? Surveys Interviews Observations Record/Document Reviews Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 87 Evaluation Data Collection Surveys Interviews Observations Record/Document Reviews Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 88 Surveys: Have a series of questions (items) with pre-determined response choices. Can include all independent items or groups of items (scales) that can be summarized. Can also include some open-ended items for write-in or clarification, Can be completed by respondents or survey administrators, Can be conducted via mail, with a captive audience, on the phone or using the internet, and through a variety of alternative strategies. Instruments are called surveys, questionnaires, assessment forms Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 89 Use Surveys: To study attitudes and perceptions. To collect self-reported assessment of changes in response to program. To collect program assessments. To collect some behavioral reports. To test knowledge. To determine changes over time. Best with big or distant groups, for sensitive information. Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 90 Evaluation Data Collection Surveys Interviews Observations Record/Document Reviews Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 91 Interviews An interview is a one-sided conversation between an interviewer and a respondent. Questions are (mostly) pre-determined, but openended. Can be structured or semi-structured. Respondents are expected to answer using their own terms. Interviews can be conducted in person, via phone, one-on-one or in groups. Focus groups are specialized group interviews. Instruments are called protocols, interview schedules or guides Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 92 Use Interviews: To study attitudes and perceptions using respondent’s own language. To collect self-reported assessment of changes in response to program. To collect program assessments. To document program implementation. To determine changes over time. Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 93 Evaluation Data Collection Surveys Interviews Observations Record/Document Reviews Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 94 Observations Observations are conducted to view and hear actual program activities so that they can be described thoroughly and carefully. Observations can be focused on programs overall or participants in programs. Users of observation reports will know what has occurred and how it has occurred. Observation data are collected in the field, where the action is, as it happens. Instruments are called protocols, guides, sometimes checklists Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 95 Use Observations: To document program implementation. To witness levels of skill/ability, program practices, behaviors. To determine changes over time. Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 96 Evaluation Data Collection Surveys Interviews Observations Record/Document Reviews Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 97 Record Review Review of program records involves accessing existing internal information or information that was collected for other purposes. Data are obtained from: a program’s own records (e.g., intake forms, program attendance) records used by other agencies (e.g., report cards; drug screening results; hospital birth data). adding questions to standard record-keeping strategies (e.g., a question for parents about program value can be added to an enrollment form). Instruments are called protocols. Use requires identification of and access to available information. Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 98 Building Evaluation Capacity Session 4 Evaluation Data Collection & Analysis Surveys and Interviews Anita M. Baker, Ed.D. Bruner Foundation Rochester, New York Evaluative Thinking • Ask important questions before decisions are made, • Systematically collect and analyze data to inform decisions, • Share results of findings and • Base responses and actions on the results of analyses (as appropriate). Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 100 Evaluation Data Collection Surveys Interviews Observations Record/Document Reviews Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 101 Surveys: Have a series of questions (items) with pre-determined response choices. Can include all independent items or groups of items (scales) that can be summarized. Can also include some open-ended items for write-in or clarification, Can be completed by respondents or survey administrators, Can be conducted via mail, with a captive audience, on the phone or using the internet, and through a variety of alternative strategies. Instruments are called surveys, questionnaires, assessment forms Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 102 Surveys Are Most Productive When They Are: Well targeted, with a narrow set of questions Used to obtain data that are otherwise hard to get. Used in conjunction with other strategies. Surveys are best used: with large numbers, for sensitive information, for groups that are hard to collect data from Most survey data are qualitative but simple quantitative analyses are often used to summarize responses. Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 103 Surveys can be administered and analyzed quickly when . . . pre-validated instruments are used sampling is simple or not required the topic is narrowly focused the numbers of questions (and respondents*) is relatively small the need for disaggregation is limited Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 104 Use Surveys To . . . study attitudes and perceptions. collect self-reported assessment of changes in response to program. collect program assessments. collect some behavioral reports. test knowledge. determine changes over time. Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 105 Benefits of Surveys Surveys can be used for a variety of reasons such as exploring ideas or getting sensitive information. Surveys can provide information about a large number and wide variety of participants. Survey analysis can be simple. Computers are not required. Results are compelling, have broad appeal and are easy to present. Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 106 Drawbacks of Surveys: Designing surveys is complicated and time consuming. The intervention effect can lead to false responses, or it can be overlooked. Broad questions and openended responses are difficult to use. Analyses and presentations can require a great deal of work. You MUST be selective. Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. ! 107 Developing/Assessing Survey Instruments 1) Identify key issues. 2) Review available literature. 3) Convert key issues into questions. 4) Determine what other data are needed. 5) Determine how questions will be ordered and formatted. 6) Have survey instrument reviewed. Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 108 For Survey Items, Remember: 1) State questions in specific terms, use appropriate language. 2) Use multiple questions to sufficiently cover a topic. 3) Avoid “double-negatives.” 4) Avoid asking multiple questions in one item. 5) Be sure response categories match the question, are exhaustive and don’t overlap. 6) Be sure to include directions and check numbering, format etc. Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 109 Types of Surveys: Mail Surveys (must have correct addresses and return instructions, must conduct tracking and follow-up). Response is typically low. Electronic Surveys (must be sure respondents have access to internet, must have a host site that is recognizable or used by respondents; must have current email addresses). Response is often better. Web + (combining mail and e-surveys). Data input required, analysis is harder. Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 110 Types of Surveys: Phone Surveys (labor intensive and require trained survey administrators, access to phone numbers, usually CATI software). Response is generally better than mail, but must establish refusal rules. Staged Surveys (trained survey administrators required, caution must be used when collecting sensitive info). Can be administered orally, multiple response options possible, response rates very high. Intercept Surveys (require trained administrators). Refusal is high. Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 111 Sampling Surveys are not always administered to every member of a group (population). Often, some members, a sample, are selected to respond. (Additional strategies in manual.) Convenience Samples. Provide useful information to estimate outcomes (e.g. 85% of respondents indicated the program had definitely helped them) Must be used cautiously, generalization limited. Random Samples. Everyone must have equal opportunity. Careful administration and aggressive follow-up needed. Generalization/prediction possible. Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 112 How Many Surveys Do you Need to Administer? Identify the population size, desired confidence and sampling error thresholds. 95% confidence with 5% error is common. With the right sample size you can be 95% confident that the answer given by respondents is within 5 percentage points of the answer if all members of the population had responded. Use this formula: n=385/(1+(385/all possible respondents)). OR Consult a probability table (see manual). Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 113 How Many Surveys Do you Need to Administer? The sample should be as large as probabilistically required. (Probability – not Percentage) If a population is smaller than 100, include them all. When a sample is comparatively large, adding cases does not increase precision. When the population size is small, relatively large proportions are required and vice versa. You must always draw a larger sample than needed to accommodate refusal. Desired sample size ÷ (1-refusal proportion) Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 114 How Can I Increase Response? Write a good survey and tailor administration to respondents. Advertise survey purpose and administration details in advance. Carefully document who receives and completes surveys. Aggressively follow-up. Send reminders. Consider using incentives. Make response easy. Remember: Non-response bias can severely limit your ability to interpret and use survey data. Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 115 Calculating Response Rates Response rate is calculated by dividing the number of returned surveys by the total number of “viable” surveys administered. Desirable response rates should be determined in advance of analysis and efforts should be made to maximize response. Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 116 Administration Plans (see apppendix12) Before you administer a survey be sure you can answer the following questions! Who and where are your target groups? Do they require assistance to answer? Which type of survey will be best to use with your target group? How often? Will the survey be anonymous or confidential? How much time will be required to respond? How will you analyze the data you expect to collect? Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 117 Administration Plans (Con’t.) What specific fielding strategy will be used? Will there be incentives? How will you track the surveys? How will you provide ample opportunities for all members of the sample to respond? What response rate is desired? Whose consent is required/desired? Will you use active or passive consent? How will you store and maintain the confidentiality of the information? Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 118 Preparing for Analysis: Developing Codebooks Unless they’re embedded, assign numbers for all response categories. Write the codes onto a copy of the survey and use for reference. It is bad practice to re-code data as you go. Prepare for entry as is. List or describe how data are to be recoded. Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 119 What Should Your Survey Analysis Plan Include? How survey items are related to evaluation overall What analytical procedures including disaggregation will be conducted with each kind of data you collect How you will present results How you will decide whether data show that targets have been exceeded, met or missed (as appropriate) How you will handle missing data Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 120 Example: Student Survey Analysis Plan 1. The percentages of all students who smoke and those who have recently started will be calculated. 2. The percentage of boys who smoke will be compared to the percentage of girls who smoke. 3. The average age of first alcohol use will be calculated from students’ responses. Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 121 Example: Continued Student Survey Analysis Plan 4. The percentage of students who provide positive (i.e., Good or Very Good) ratings for the smoking prevention program will be calculated. Answers for self-reported nonsmokers will be compared to self-reported smokers. 5. The distribution of scores on the likelihood of addiction scale will be determined. Only valid percents will be used, items missed by more than 10% of respondents will not be used. Meeting the target means ±5 percentage points. Far exceeding or missing = +15 percentage points. Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 122 Common Factors That Can Influence Responses Participant Characteristics • • • • • Age group, gender, race/ethnicity Educational level or type Household income group, household composition Status (e.g., disabled/not, smoker/non) Degree of difficulty of the participant’s situation Location of program • Political or geographic boundaries • Program sites • Characteristics of location (e.g., distressed/not) Program Experience (type or amount or history) Where appropriate, disaggregate by one or more of these! Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 123 Survey Result Example Disaggregated Data Peer Study Group Total Yes n=232 No n=247 N=479 Reported struggling to maintain grades 36% 58% 47% Are planning to enroll for the sophomore year at this school 89% 72% 80% % of 2005 Freshman who . . . Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 124 Survey Result Example Comparison of Site Outcomes: MS/JHS Only CAP Site 1 CAP Site 2 CAP Site 3 100% 80% 60% 40% 20% 0% an Pl H h ig h ig to C ue in t on m ra og pr le du he n io t ac t os st f is m Sc s de ge lle Co e er t In t Sa a s as Cl ed nd tte H A d d re te si Vi lte T ra G A PS ed in e th a nt ai A M ok To s Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 125 Survey Result Example Table 3: Relationships between ATOD use and other factors among 9th graders Ever Used Never Used ATOD ATOD Academically and socially attached to school 35% 49% Have post-secondary aspirations 68% 73% Are passing most classes 86% 94% Were sent to the office during last 2 months 23% 10% Describe their health as excellent 30% 42% Felt unhappy, sad or depressed recently 32% 12% Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 126 Evaluation Data Collection Surveys Interviews Observations Record/Document Reviews Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 127 Interviews An interview is a one-sided conversation between an interviewer and a respondent. Questions are (mostly) pre-determined, but openended. Can be structured or semi-structured. Respondents are expected to answer using their own terms. Interviews can be conducted in person, via phone, one-on-one or in groups. Focus groups are specialized group interviews. Instruments are called protocols, interview schedules or guides Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 128 Use Interviews: To study attitudes and perceptions using respondent’s own language. To collect self-reported assessment of changes in response to program. To collect program assessments. To document program implementation. To determine changes over time. Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 129 Interviews: Methodological Decisions What type of interview should you conduct? (see pg. 28) Unstructured Semi-structured Structured Intercept What should you ask? How will you word and sequence the questions? What time frame will you use (past, present, future, mixed)? Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 130 Interviews: More About Methodological Decisions How much detail and how long to conduct? Who are respondents? (Is translation necessary? How many interviews, on what schedule? Will the interviews be conducted in-person, by phone, on-or-off site? Are group interviews possible/useful? Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 131 Conducting and Recording Interviews: Before Clarify purpose for the interview. Specify answers to the methodological decisions. Select potential respondents – sampling. Collect background information about respondents. Develop a specific protocol to guide your interview. Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 132 Conducting and Recording Interviews: During Use the protocol (device) to record responses. Use probes and follow-up questions as necessary for depth and detail. Ask singular questions. Ask clear and truly open-ended questions. Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 133 Conducting and Recording Interviews: After Review interview responses, clarify notes, decide about transcription. Record observations about the interview. Evaluate how it went and determine follow-up needs. Identify and summarize some key findings. Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 134 Tips for Effective Interviewing Communicate clearly about what information is desired, why it’s important, what will happen to it. Remember to ask single questions and use clear and appropriate language. Avoid leading questions. Check (or summarize) occasionally. Let the respondent know how the interview is going, how much longer, etc. Understand the difference between a depth interview and an interrogation. Observe while interviewing. Practice Interviewing – Develop Your Skills! Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 135 More Tips Recognize when the respondent is not clearly answering and press for a full response. Maintain control of the interview and neutrality toward the content of response. Treat the respondent with respect. (Don’t share your opinions or knowledge. Don’t interrupt unless the interview is out of hand). Practice Interviewing – Develop Your Skills! Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 136 Analyzing Interview Data 1) Read/review completed sets of interviews. 2) Record general summaries 3) Where appropriate, encode responses. 4) Summarize coded data 5) Pull quotes to illustrate findings. (see pg 30 for examples) Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 137 What Happens After Data are Collected? 1. Data are analyzed, results are summarized. 2. Findings must be converted into a format that can be shared with others. 3. Action steps should be developed from findings Step 3 moves evaluation from perfunctory compliance into the realm of usefulness. “Now that we know _____ we will do _____.” Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 138 Increasing Rigor in Program Evaluation Mixed methodologies Multiple sources of data Multiple points in time Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 139 Building Evaluation Capacity Session 5 Evaluation Data Collection & Analysis Observation and Record Review Anita M. Baker, Ed.D. Bruner Foundation Rochester, New York What strategies are used to collect data about indicators? Surveys Interviews Observations Record/Document Reviews Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 141 Evaluation Data Collection Surveys Interviews Observations Record/Document Reviews Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 142 Observations Observations are conducted to view and hear actual program activities so that they can be described thoroughly and carefully. Observations can be focused on programs overall or participants in programs. Users of observation reports will know what has occurred and how it has occurred. Observation data are collected in the field, where the action is, as it happens. Instruments are called protocols, guides, sometimes checklists Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 143 Use Observations: To document program implementation. To witness levels of skill/ability, program practices, behaviors. To determine changes over time. Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 144 Trained Observers Can: see things that may escape awareness of others learn about things that others may be unwilling or unable to talk about move beyond the selective perceptions of others present multiple perspectives Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 145 Other Advantages the observer’s knowledge and direct experience can be used as resources to aid in assessment feelings of the observer become part of the observation data OBSERVER’S REACTIONS are data, but they MUST BE KEPT SEPARATE Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 146 Observations: Methodological Decisions What should be observed and how will you structure your protocol? (individual, event, setting, practice) How will you choose what to see? Will you ask for a “performance” or just attend a regular session, or both? Strive for “typical-ness.” Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 147 Observations: Methodological Decisions Will your presence be known, or unannounced? Who should know? How much will you disclose about the purpose of your observation? How much detail will you seek? (checklist vs. comprehensive) How long and how often will the observations be? Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 148 Conducting and Recording Observations: Before Clarify the purpose for conducting the observation Specify the methodological decisions you have made Collect background information about the subject (if possible/necessary) Develop a specific protocol to guide your observation Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 149 Conducting and Recording Observations: During Use the protocol to guide your observation and record observation data BE DESCRIPTIVE (keep observer impressions separate from descriptions of actual events) Inquire about the “typical-ness” of the session/event. Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 150 Conducting and Recording Observations: After Review observation notes and make clarifications where necessary. clarify abbreviations elaborate on details transcribe if feasible or appropriate Evaluate results of the observation. Record whether: the session went well, the focus was covered, there were any barriers to observation there is a need for follow-up Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 151 Observation Protocols Comprehensive Setting Beginning, ending and chronology of events Interactions Decisions Nonverbal behaviors Program activities and participant behaviors, response of participants Checklist – “best” or expected practices Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 152 Analyzing Observation Data Make summary statements about trends in your observations Every time we visited the program, the majority of the children were involved in a literacy development activity such as reading, illustrating a story they had read or written, practicing reading aloud. Include “snippets” or excerpts from field notes to illustrate summary points (see manual pp 38-39) Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 153 Analyzed Observation Data Many different types of arts activities were undertaken, and personal development was either delivered directly or integrated with arts activities. Of the 57 different combinations of programming at the 10 sites, only 3 included activities that were not wholly successful with their target groups, 2 of those because of mismatch between instructor and the participant group. At all sites, ongoing projects were underway and examples of participant work were readily visible. Teaching artists were demonstrating skills, giving youth opportunities to try the skills, and providing one-on-one assistance as needed. Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 154 Evaluation Data Collection Surveys Interviews Observations Record/Document Reviews Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 155 Record Review Review of program records involves accessing existing internal information or information that was collected for other purposes. Data are obtained from: a program’s own records (e.g., intake forms, program attendance) records used by other agencies (e.g., report cards; drug screening results; hospital birth data). adding questions to standard record-keeping strategies (e.g., a question for parents about program value can be added to an enrollment form). Instruments are called protocols. Use requires identification of and access to available information. Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 156 Use Record Reviews: To collect some behavioral reports. To test knowledge To verify self-reported data. Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. To determine changes over time. 157 Analyzing/Using Record Review Data Findings from record review data are usually determined through secondary analysis. Example: Attendance data are regularly collected for a program to inform routine program operations. Attendance records are summarized quarterly or annually to inform other stakeholders such as funders about program use. Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 158 Analyzing/Using Record Review Data Results of record reviews are typically arrayed in tables or summarized in profiles, or “bullet lists” as frequencies or proportions, or averages (see pg. 16, appendix 10 in the Participatory Evaluation Essentials Guide). Like observation data, record review data can be both descriptive and/or evaluative. -- See pg 16. Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 159 Analyzing/Using Record Review Data Record review data are commonly combined for multi-variate analyses with other evaluation data to determine relationships Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 160 Collecting Record Review Data Review existing data collection forms (suggest modifications or use of new forms if possible). Develop a code book or at least a data element list keyed to data collection forms. Develop a “database” for record review data. Develop an analysis plan with mock tables for record review data. Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 161 Record Review Data: Example ASAP Participant Outcomes New York Number Enrollment Goal 188 Enrollment Actual 152 Trn. Completion Goal 97 Trn. Completion Actual Placement Actual (30+) Placement Actual (180+) In-field placement 87 41 83 77 Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. % Boston Number % 112 81% 94 84% 48 89% 48% 97% 93% 39 26 37 36 81% 59% 84% 97% 162 Record Review Data: Example Outcome: Delivering Healthy Babies In Program Number On Waiting List % Number % Babies Born 18 Born Healthy* 13 72% 14 64% Not Born Healthy* 5 28% 8 36% 22 *The indicator of a healthy baby is birthweight above 5.5 pounds AND Apgar score 7 Or Above. Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 163 Record Review Data: Example Average Pre and Post Test Scores for Youth Enrolled in Summer Learning Camps Average Scores Pre Test Post Test Difference Reading 22.7 (64%) 25.2 (72%) + 2.5 Math 29.9 (85%) 29.7 (85%) -0.2 Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 164 What Happens After Data are Collected? 1. Data are analyzed, results are summarized. 2. Findings must be converted into a format that can be shared with others. 3. Action steps should be developed from findings Step 3 moves evaluation from perfunctory compliance into the realm of usefulness. “Now that we know _____ we will do _____.” Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 165 Increasing Rigor in Program Evaluation Mixed methodologies Multiple sources of data Multiple points in time Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 166 Building Evaluation Capacity Session 6 Designing Evaluations Putting it All Together Anita M. Baker, Ed.D. Bruner Foundation Rochester, New York Good Evaluation Designs Include the Following Summary Information about the program The questions to be addressed by the evaluation The data collection strategies that will be used The individuals who will undertake the activities When the activities will be conducted The products of the evaluation (who will receive them and how they should be used) Projected costs to do the evaluation Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 168 Increasing Rigor in Program Evaluation Mixed methodologies Multiple sources of data Multiple points in time Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 169 What else must you think about? Data Collection Management Identify data sources Select data collection methods Develop and/or test instruments and procedures Develop plans for entering and managing the data Train data collectors Plan for analysis Plan to monitor the data collection system Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 170 Thinking about . . . . Data Collection Instruments • Who will you collect data about? Clients, caregivers, other service providers working with clients, staff, some other group? Who are considered participants of your program? Be sure to clearly specify your eval. target population. • What instruments do you need? Surveys, interview guides, observation checklists and/or protocols, record extraction protocols? • Are there any pre-tested instruments (e.g., scales for measuring human conditions and attitudes)? – If not, how will you confirm validity? Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 171 Thinking about . . . . Data Collection Instruments Keeping in mind things like cultural sensitivity, language and expression: • Are the instruments you plan to use appropriate for the group you are planning to use them with? • Will responses be anonymous or confidential? • How will you analyze data from instruments you choose? Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 172 Thinking about . . . . Data Collection Procedures • What are your timelines for data collection? – When will you administer surveys, conduct interviews, etc. ? – Are pre/post strategies needed? Doable? • When do you need data? – Is this the same time that data collectors and subjects are available? – What outcomes are expected by the time data collection is planned? i.e., is this the proper timeframe? • What is required for data collection approval? – – – – Institutional review? Active consent? Passive consent? Informed consent? Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 173 Thinking about . . . . Data Entry and Management • How will you store and maintain the information you collect? – How much data is expected and in what form? – What procedures are necessary to ensure confidentiality? – Where will the data reside? • How will you handle data entry? – Do you have specialty software or can you use readily available programs like Excel to help support your data entry? – Who will actually enter the data and where will it be entered? Are there training needs? Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 174 Thinking about . . . . Data Collector Training • Who will collect the data? Staff within a program, staff from another program, other agency staff, clients from another program (e.g., youth), volunteers? • What training do data collectors need? – Can they administer surveys? – Do they know how to conduct interviews? – Have they been trained as observers for this data collection? – Do they have access to and knowledge about records? Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 175 Thinking about . . . . Data Analysis • How will you analyze the data you collect? – How will you handle quantitative data? e.g., frequencies, averages, ranges, distributions? Do you need tables and graphs? Do you know how to make them? – How will you handle qualitative data, e.g., quotes, “snippets,” numerical summaries? – What will you do about missing data? – What influencing factors should you consider? What disaggregation is needed? • Who (staff, volunteers, consultants) will conduct the analysis and how long will it take? Will they need some additional training? • Are there any additional costs associated with data analysis? Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 176 What are the Components of a Strong Evaluation Report? * Subject program description. * Clear statement about the evaluation questions and the purpose of the evaluation. * Description of actual data collection methods used. * Summary of key findings (including tables, graphs, vignettes, quotes, etc. * Discussion or explanation of the meaning and importance of key findings * Suggested Action Steps * Next Steps (for the program and the evaluation). * Issues for Further Consideration (loose ends) Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 177 Additional Reporting Tips Findings can be communicated in many forms. * * * * brief memos powerpoint presentations oral reports formal evaluation report is most common Think about internal and external reporting. Plan for multiple reports. Before you start writing, be sure to develop an outline and pass it by some stakeholders. If you’re commissioning an evaluation report, ask to see a report outline in advance. If you are reviewing others’ evaluation reports, don’t assume they are valuable just because they are in a final form. Review carefully for the important components and meaningfulness. Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 178 Projecting Level of Effort LOE projections are often summarized in a table or spreadsheet. To estimate labor and time: • List all evaluation tasks • Determine who will conduct each task • Estimate time required to complete each task (including pre-training), in day or half-day increments (see page 42 in Participatory Evaluation Essentials) Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 179 Projecting Timelines Timelines can be constructed separately or embedded in an LOE chart (see example pp. 44 – 45 Participatory Evaluation Essentials). To project timelines: • Assign dates to your level of effort, working backward from overall timeline requirements. • Be sure the number of days required for a task and when it must be completed are in sync and feasible. • Check to make sure evaluation calendar is in alignment with program calendar. Don’t plan to do a lot of data collecting around program holidays Don’t expect to collect data only between 9 and 5 Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 180 Budgeting and Paying for Evaluation • Usually the cost to do good evaluation is equivalent to about 10 – 15% of the costs to operate the program effectively. • Most of the funds for evaluation pay for the professional time of those who develop designs and tools, collect data, analyze data, summarize and present findings. • Other expense include overhead and direct costs associated with the evaluation (e.g., supplies, computer maintenance, communication, software) Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 181 Projecting Budgets • Determine rates for all “staff” to the project. • Calculate total labor costs by multiplying LOE totals by “staff” rates. • Estimate other direct costs (ODC) such as copying, mail/delivery, telephone use and facilities. • Estimate any travel costs. • Calculate the subtotal of direct costs including labor (fringe where appropriate), ODC and travel. • Estimate additional indirect (overhead) costs, where appropriate, as a percentage applied to the direct costs. • Apply any other fees where appropriate • Sum all project costs to determine total cost of project. • Establish a payment schedule, billing system and deliverables. Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 182 Things to Avoid when Budgeting and Paying for Evaluation • It’s bad practice to assume there is a standard, fixed evaluation cost regardless of program size or complexity. • It is dangerous to fund an evaluation project that does not clarify how evaluation funds will be used. Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 183 Budgeting and Paying for Evaluation There are two ways to project evaluation costs: Identify a reasonable total amount of funds dedicated for evaluation and then develop the best evaluation design given those resource requirements. Develop the best evaluation design for the subject program, and then estimate the costs associated with implementing the design. NEGOTIATE design changes if costs exceed available funds. Bruner Foundation Rochester, New York Anita M. Baker, Ed.D. 184 These materials are for the benefit of any 501c3 organization. They MAY be used in whole or in part provided that credit is given to the Bruner Foundation. They may NOT be sold or redistributed in whole or part for a profit. Copyright © by the Bruner Foundation 2007 * Please see the notes attached to the first slide for further information about how to use the available materials. Bruner Foundation Rochester, New York Anita M. Baker, Ed.D.