Outcome Measurement and Evaluation For Nonprofit Learning Point S Overview S Introduction to Evaluation and Measurement S Why Should You Measure Outcomes? S Creating a Framework for Identifying Appropriate Outcomes for Your Program – the Logic Model S How to Measure Outcomes S Brief Review of Results-Based Accountability Gauging Your Needs S What is the mission of your organization? S What services do you provide? S Why are you interested in this class? S What do you hope to take away today? S Who do you plan to share it with? 3 Seize the Opportunity Continual improvement is an unending journey. - Lloyd Dobens 4 Balanced Perspectives I don't have a particular recommendation other than that we base decisions on as much hard data as possible. We need to carefully look at all the options and all their ramifications in making our decisions. - Dorothy Denning 5 What is Evaluation? S Program evaluation is carefully collecting and analyzing information about a program or some aspect of a program in order to make necessary decisions. S The type of evaluation you undertake to improve your program depends on what you want to learn. 6 Evaluation and Measurement S Measurement techniques help us collect useful feedback. The purpose of measurement is to raise our awareness, presenting opportunities for improvement. S Data results from our efforts to measure our program services, participants, and results. S Information results from processing data. “The height of insanity is doing things the same way and hoping for a different result.” 7 Myths About Evaluation S Evaluation is a useless activity that generates lots of boring data with useless conclusions. S Evaluation is about proving the success or failure of a program. S Evaluation is a highly unique and complex process. S Evaluation is only for developing organizations. S Nobody cares about program evaluations. 8 What Does Evaluation Help Us Do? 9 What Does Evaluation Help Us Do? S Determine where we actually are 10 What Does Evaluation Help Us Do? S Determine where we actually are S Set goals for where we want to be 11 What Does Evaluation Help Us Do? S Determine where we actually are S Set goals for where we want to be S Develop a plan for how to get there 12 What Does Evaluation Help Us Do? S Determine where we actually are S Set goals for where we want to be S Develop a plan for how to get there S Track progress on implementing the plan 13 What Does Evaluation Help Us Do? S Determine where we actually are S Set goals for where we want to be S Develop a plan for how to get there S Track progress on implementing the plan S Change the plan, if necessary 14 Why Measure Outcomes? S Gather information that will be useful for decision making/program development S Identify strong and weak components of programs S Assist in demonstrating a program’s value and effectiveness S Help justify existence S Guide budget development and resource allocations Roadblock to Meaningful Evaluation S Key Problem: Priority measurement issues not identified S What areas are most in need of improvement? S What benefits do we need to demonstrate? S What data do we have? Is it too much? Not enough? Does the data really tell us what we want to know? 16 Know Your Audience… Exercise Possible Audience What Do They Already Know? What Do They Need to Know? 17 How Will They Use the Data? Develop Balanced Measures S Include both process and outcomes measures S Process –activity/output/efficiency S Number and type of volunteer hours provided, volunteers trained, workshops held, applications processed, etc. S Outcomes – effectiveness/results/impact S Can include client and volunteer satisfaction, change in clients’ level of awareness or behaviors, program referrals from participants, number of new processes developed, etc. S Examples: S Avg. contact time with participants (process) S Avg. number of participants who demonstrate increased knowledge of effective parenting skills (outcomes) 18 Develop Balanced Measures S Address successes as well as challenges S Successes –accomplishments S Percentage of clients who recommend the service to others, percentage of students who attained a GED, etc. S Challenges – areas for improvement S Error rates, complaints, terminations, etc. 19 Make Your Goals Measurable Examples S Process: S Increase the number of families served to 50 by September 30, 2007 S Outcomes: S Achieve a 90% exceptional rating on client satisfaction surveys within 6 months S Increase reported use of substance abuse refusal skills by 20% after 10 weeks S 75% of all participants will achieve independent living status by January 31, 2007 20 Make Your Goals Measurable Increase reported use of substance abuse refusal skills by 20% after 10 weeks Audience: Who do you plan to impact? Indicator: Target: Timeline: Baseline: Use of SA refusal skills 20% increase 10 weeks Pre-intervention use of refusal skills 21 Be Realistic S Collect data that is actionable S Don’t measure A and hope for B S Measuring the hours a counselor spends with a client is not a measure of how that contact has impacted the client. 22 Creating a Framework for Identifying Appropriate Outcomes - the Logic Model What is a logic model? A logic model is a simple description of how your program works that shows the linkages between: SProblem you are attempting to address SProgram components SProgram activities SOutcomes (both short- and long-term) Why Should You Develop a Logic Model for Your Program? S Useful for designing/developing your program S Provides a useful framework for evaluating your program S Provides a focal point for stakeholders, requiring them to work together to identify the components of the model and to think about expected outcomes. Existing Problem S Identification of risk factors or needs in the community S Availability of program resources in the community S Availability of fiscal resources S Type of community partnerships that already exist in the community Program Components Components are closely related groups of: S Activities in your program that are logically linked to the conditions identified as part of the program planning process. S The number of components depends on the size of your program and how you administer it. For a large program, there could be several components in the logic model. Smaller programs might consist of just one component. Examples of Program Components S Parent component – a group of program activities conducted with parents S Family component – a group of program activities conducted with the entire family S Youth component – a group of activities conducted with children/youth in the program Program Activities S Activities are the services a program provides for its participants and the means through which the desired outcomes will be achieved. S Activities are logically connected to the identified program components. Examples of Program Activities S Within a parent component, activities might include a 12- week session of parenting classes and a parent support group held twice a week S Within a family component, activities might include family counseling sessions and family recreational activities held twice per week S Within a youth component, activities might include youth recreational activities and youth counseling sessions twice per week Program Outcomes S Outcomes are benefits or changes for program participants during or after their involvement with a program that flow logically from the precipitating program activity. S Outcomes can be short-term or long-term S Outcomes are not process measures Short-Term Outcomes Short-term outcomes should: S Flow logically from the precipitating activity; S Be relevant to the program’s overall goal; S Be reasonable to achieve; and S Represent achievements or improvements on the part of program participants that occur either during or immediately after program participation ends. Examples of Short-Term Outcomes S Parents in parenting class increase knowledge of parenting skills S Youth attending a drop-out prevention program improve school attendance Long-Term Outcomes Long-term outcomes should: S Have a logical connection to the short-term outcomes; S Represent benefits or changes for program participants over an extended period of time; S Be collected during a follow-up period with program participants. Examples of Long-Term Outcomes S Parents completing a parenting class improve parenting skills, including ability to supervise and discipline their children S Youth participating in a drop-out prevention program stay in school and graduate S Youth completing a diversion program are not re-adjudicated one year after program participation ends What Information Should be Included in a Program Outcome? Outcomes are simple goal statements that identify the following components: S The target population S What benefit or change is occurring S How the benefit or change will be measured What is the Target Population? The target population is the individuals, groups, or communities to whom the program’s services are directed. This is your program’s priority population or the intended “audience” for your program. Examples might be: S Males between the ages of 14 and 16 S Parents with children between the ages of 3 and 5 S Families in which abuse is occurring What Benefit or Change is Occurring? S Knowledge S Attitude S Skills S Values S Behavior How are Outcomes Measured? S Setting Outcome Targets S Defining Amount of Change in Target Population S Describe Participant Eligibility to Meet Outcome S Selection of Instruments or Data Collection Tools Setting Outcome Targets S Outcome targets are the numerical objectives for the number and percent of participants who will achieve the desired outcome in the next reporting period. S Examples include: S 75% of participating students will improve grades S 50% of families will improve family functioning Defining the Amount of Change in Target Population S The amount of change describes specifically how much impact you expect the program to achieve. S The amount of change should ideally be measurable. S For example: S 75% of participating students will improve their grade point average by one letter grade Participant Eligibility S What level of program participation is required before you would expect to achieve the desired outcomes? S Allows you to assess differences in outcomes for clients who follow the base program requirements vs.those who do not. S For example: S 75% of students who attend 30 of 40 program sessions will improve their grade point average by one letter grade Selection of Instruments or Data Collection Tools Things to Consider When Selecting Instruments or Data Collection Tools: S Purpose and Feasibility S Administration of Instrument S Utility and suitability S Psychometric qualities Purpose and Feasibility S Is the tool’s purpose consistent with your needs? S Does the instrument measure issues that relate directly to your desired outcomes? S Are there clear directions for administration? S Is instrument easy to administer and score? S Is training needed prior to administration? S Does program have adequate resources to cover cost of instrument, scoring, and training, if applicable? Administration of Instrument S Number of items S Length of time to complete S Method of administration (self-report, rating by trained clinician, other) S Recommended administration schedule Utility and Suitability of Instrument S How will data be used? S What reading and writing levels are needed by respondents? S Is tool culturally sensitive? S Is tool available in other languages? S Does tool have good clinical fit with other instruments program is using? Psychometric Qualities of Instrument S Tested groups S Reliability S Validity Specific Ways to Measure S Surveys S Personal interviews S Focus groups S Other data collections tools 47 Who Can Be Helpful? S Depends on the purpose – who can answer your questions? S If not apparent, ask for guidance from key informants. S If potential population is large, generate a list of candidates and key characteristics then randomly select. 48 What is a Survey? A quantitative research project in which a relatively large number of people are queried, each being asked a standard set of questions, posed in the same way each time. 49 Why Use Surveys? S To collect specific, standardized data across respondents. S To collect data across many people as efficiently as possible. S To increase understanding. S To collect data quickly, even across multiple groups. S Open-ended quotes may add impact and credibility. 50 Types of Surveys Method S Snail Mail S Web-Based S Phone S Blended Approach Question Types S Open-ended S Closed-ended S Combination 51 The Survey Tool DO: S Keep the survey as short as possible. S Balance white space and length. S Keep questions neutral. S Be specific in your wording. S Pretest your questions. S Address only one construct per question. S Use a blend of closed and open-ended questions, if reasonable. S Mix types of questions throughout. S Provide an opportunity for additional comments. 52 The Survey Tool DON’T: S Use leading questions. S Make your questions too complicated. S Ask controversial questions first. S Mix reference points. 53 How to Enhance Participation DO: S Provide a reason for the survey. S Use clear instructions. S Explain confidentiality protections. S Give a realistic sense of the time involved. S Track respondents. S Use reminders. 54 How to Enhance Participation DO: S Provide your contact information. S Request contact information from the respondent, if appropriate. S Establish an end date. 55 Using Survey Data S Prepare your analysis plan. 56 What is an Interview? An interview is the collection of data by asking people questions and following up or probing their answers, usually conducted face-to-face or by phone. 57 Why Use Interviews? S As an exploratory first step to creating quantitative tools. S Enhance understanding of interesting findings which emerged from other processes. S Real-world quotes may add impact and credibility. S To collect specific, standardized data across respondents. 58 Types of Interviews Method S Face-to-face S Telephone Style S Conversational: go with the flow S Interview guide 59 Conversational Approach S Questions emerge through conversation S Highly individualized S May provide greater insights S Requires interviewer with strong content and interpersonal skills S Not systematic S Can be more difficult to analyze 60 Interview Guide Approach S Outline of topics exists, but wording and order can vary S Topics covered is more systematic S Interview setting remains fairly informal S Also required considerably skilled interviewer S Inhibits spontaneous discussion of new topics 61 Open-Ended Structured Approach S Strict script for question, but free-response format for answers S Most efficient of qualitative techniques S Reduces bias and guides less experienced interviewers 62 The Interview Tool DO: S Keep the interview as short as possible. S Keep questions neutral. S Be specific in your wording. S Pretest your questions. S Address only one construct per question. S Use a blend of closed and open-ended questions, if reasonable. S Provide an opportunity for additional comments. 63 The Interview Tool DON’T: S Use leading questions. S Make your questions too complicated. S Ask controversial questions first. S Mix reference points. 64 Recruitment Techniques DO: S Provide a reason for the interview. S Give a realistic sense of the time involved. S Explain how your interview candidate fits into the bigger goal. 65 Recruitment Techniques DON’T: S Be vague about why you need the interview. S Candy coat the commitment. 66 Preparing for the Interview DO: S Plan an appropriate interview location. S Exchange contact information with the interviewee. S Prepare to capture information. S Create response cards, if needed. 67 Preparing for the Interview DON’T: S Choose a location with lots of distractions. S Forget to take the interviewees phone number with you! 68 Building Rapport DO: S Be on time. S Review the purpose of the interview. S Discuss how the results will be used. S Reiterate confidentiality protections. S Acknowledge the time frame you have set aside. S Provide your contact information on paper. S Ask if they have questions before you begin. 69 Building Rapport DON’T: S Be a robot! 70 General Tips S Don’t be judgmental! S Use appropriate non-verbal cues. S Attend to non-verbal cues. S Repeat and clarify when needed. S Practice active listening. S Let the interviewee know when you are switching gears. S Keep the interviewee focused. S Don’t make assumptions. 71 What is a Focus Group? A focus group is a group discussion. Participants are brought together in a neutral location for the specific purpose of discussing an issue or responding to ideas or materials of interest. 72 Why Use Focus Groups? S To increase understanding. S As an exploratory first step to creating quantitative tools. S Enhance understanding of interesting findings which emerged from other processes. S Real-world quotes and rich feedback may add impact and credibility. 73 Types of Focus Groups Method S Exploratory S Issue-Focused 74 Focus Group Model DO: S Plan on a 2 hour session S Secure 8-12 participants and a few alternates S Identify 5-7 key questions beforehand S Provide background information to group. S Keep questions neutral. S Record all input. S Have SMEs available to answer questions. S Use an objective facilitator. 75 The Focus Group Session DON’T: S Use leading questions. S Accept too many participants. S Try to squeeze in too many questions. S Forget to document results. 76 Recruitment Techniques DO: S Provide a reason for the focus group. S Give a realistic sense of the time involved. S Explain how your focus group candidate fits into the bigger goal. S Establish a potential participant list. S Invite participants by letter, followed by phone calls. S Consider offering incentives. 77 Preparing for the Session DO: S Secure a roomy, centrally-located venue S Select a facilitator S Exchange contact information with the participants S Prepare background information S Prepare to capture information 78 Building Rapport DO: S Be on time. S Review the purpose of the focus group. S Discuss how the results will be used. S Reiterate confidentiality protections. S Acknowledge the time frame you have set aside. S Provide your contact information on paper. S Ask if they have questions before you begin. 79 Other Tools S Observations S Checklists (process documentation) S Attendance logs S Case records S Existing databases 80 Data Management S Develop a plan to maintain program and outcome information S Excel can be a useful tool What is Results-Based Accountability (RBA)? S Draws a clear distinction between community accountability for population results vs. manager accountability for program client results. S Addresses accountability from the highest level view across systems and across communities to accountability for the smallest program in the bureaucracy, and everything in between. Results-Based Accountability Implementation Guide (www.raguide.org) 82 How Does RBA Fit Into the Levels of Outcomes? Outcomes Level Community Systems Collaborative Organizational Program Client Description Collected across organizations/programs to address changes in community-level needs or conditions Collected across and/or outside organizations to assess changes in capacity or systems Collected, across programs or organizations, to assess results for clients of collaborative service efforts Collected, across multiple programs, to assess results for clients of an organization Collected, across multiple clients, to assess results in a group of individuals who have received a service or set of services Collected to assess changes in an individual client or a group of clients as a result of a provided service Often Driven By: Community Collaboratives, Funding Collaboratives (Results-Based Accountability) Systems Change Agents, Organizations, Funders Collaborative Partners, Organizational Leadership, Funders Organizational Leadership, Funders Program Staff, Organizational Leadership, Funders Program staff, Organizational Leadership, Funders 83 Other Supporting Data: •Administrative •Implementation (progress, barriers, integrity, compliance) •Workload (hours invested) •Output/Process (#s served) •Financial (costs) •Program Development (satisfaction, attendance) RBA – Population Accountability S Population accountability is accountability: S BY THE COMMUNITY (city, county or state)… S TO THE COMMUNITY (city, county or state)… S FOR THE WELL-BEING (RESULTS) OF A POPULATION (children, adults, families, all citizens...). Results-Based Accountability Implementation Guide (www.raguide.org) 84 RBA – Performance Accountability S Performance accountability is accountability: S BY THE MANAGERS (of the program, agency or service system)… S TO THE CUSTOMERS (and other stakeholders)… S FOR THE PERFORMANCE OF THE PROGRAM (agency or service system). Results-Based Accountability Implementation Guide (www.raguide.org) 85 Other Key Concepts of RBA S Results Decision-Making S Results-Based Budgeting S Language of Accountability Results-Based Accountability Implementation Guide (www.raguide.org) 86 The Community Foundation’s Implementation of RBA S Long-Term Goals S Performance Measures: S How Much Do You Do? S How Well Do You Do? S Will Anyone Be Better Off ? S Was Anyone Better Off Last Year? S Data Collection Methods 87 Parting Tips S Plan ahead, plan ahead, plan ahead S Focus on continuous improvement S Include outcomes/results measures S Keep it simple 88 Questions? Please contact: Trina Willard Principal Knowledge Advisory Group 804-564-6969 Trina@KnowledgeAdvisoryGroup.com 89