A Premier Evaluation Agency: USAID Progress in Capacity Building Cindy Clapp-Wincek Director, Office of Learning, Evaluation & Research Bureau for Policy, Planning and Learning USAID June 19, 2013 1 Guidance 2 Technical Support • TA and advice from LER staff • Program Cycle Service Center • ProgramNet and Learning Lab http://programnet.usaid.gov/ www.usaidlearninglab.org 3 Systems • Special Collection on Evaluation http://dec.usaid.gov • Evaluation Showcase www.usaid.gov/evaluation • Transparency – Evaluation Registry tracks evaluations and budgets – Waivers for public posting on DEC 4 Progress Year Number of DEC Documents Verified as Evaluations Number of Verified Evaluations Coded Included in the Meta-Evaluation 2009 112 73 2010 142 85 2011 154 89 2012 165 93 Combined 573 340 Statistical Characteristics of Samples Drawn Separately for Each Year of the Meta Evaluation For each sample year: Confidence Level: 85% Margin of Error: +/- 5% Confidence Level: > 99% Margin of Error: +/- 5% 5 3 The Long March Progress 6 3 7 3 Planned and Actual Use of Evaluation Data Collection Methods Evaluation Described Plans to Use the Method Report Review Found Evidence of Use of the Method Difference between Plan to Use and Actual Use Percentage of Evaluations that Demonstrated Use of the Method USAID Performance Data 243 285 +42 (117%) 84% Document Review 252 274 +22 (109%) 81% Key Informant Interviews 261 245 -16 (94%) 72% Individual Interviews 187 185 -2 (99%) 54% Unstructured Observation 156 152 -4 (97%) 45% Survey 143 118 -25 (83%) 35% Focus Group 147 100 -47 (68%) 29% Structured Observation 24 26 +2 (108%) 8% Group Interview 64 32 -32 (50%) 9% Instruments (e.g., scale) 9 11 +2 (122%) 3% Community Interview 5 3 -2 (60%) 1% Collection Methods 38 Elements an Evaluation Client Determines Scope of the evaluation – single or multiple projects or programs Timing of the evaluation – during implementation, towards the end of a project, and evaluation schedule Management purpose – improve performance, generate lessons Type of evaluation sought – performance, impact Evaluation questions – number and types Team composition – external evaluation team leader, evaluation specialist, local evaluators Identification of deliverables, and the transmission of Agency evaluation quality standards Duration – number of weeks or months Evaluation budget Evaluation quality control activities, including evaluation product reviews Elements an Evaluation Team Provides Executive Summary – degree to which it accurately mirrors most critical elements of the report Presentation of Project or Program Background – completeness from a reader’s perspective Description of the Project or Program’s “Theory of Change” – development hypotheses Presentation of the Evaluation Questions – consistency with SOW, completeness Description of the Data Collection and Analysis Methods Used – specificity, links to questions Description of the Study Limitation Findings, Conclusions and Recommendations – clear distinctions among them, logical flow Annexes -- completeness 9 3