Impact assessment Training 10/11th September 2013 Handouts Handout 1: Overview of Impact Assessment Definitions: Impact Assessment can be defined as “the systematic analysis of significant and/or lasting change – positive or negative, intended or not – in the lives of target groups, brought about by a given action or a series of actions”. The guidelines are designed to enable you to carry out this systematic analysis. Impact Assessment – So What? Impact assessment is designed to ask (and answer) the “SO WHAT” questions: we have completed our project/programme successfully: So what has actually changed? For whom? How significant have these changes been for different target groups? How did these changes come about? What are the factors contributing to them? What, if anything, did our programme contribute to these changes? So what should we do differently next time? Purpose: The overall goal for all those of us who work in development should be to improve the quality of life for those men, women, girls and boys in the communities where we are work. The most important reason therefore, for assessing the impact of our efforts is to learn about what works, what difference are we making and to continually try to improve our effectiveness. Additionally, we need to be able to report on impact for our donors (although donors mostly only focus on outcomes or, at best, expected impacts in relation to log frames). We are accountable to all of our stakeholders, especially the communities with whom we work : we should be working together with them to identify changes that they want to see in their lives; and to monitor and assess how well our projects and programmes are contributing to these identified changes Lastly, but not least, evidence of impact is a very powerful tool for advocacy (for example, evidence of numbers of children whose lives have improved as a result of a change of law or policy); and for inspiration. Impact assessment 10/11th September 2013 Handouts Page 2 Handout 2: Some Key Challenges to Assessing Impact While many organisations have successfully developed processes to monitor and evaluate their projects and programmes (at the level of outputs and outcomes), the majority fail to look beyond their individual projects or programmes to really assess the difference they are making in people’s lives. They struggle to assess the impact that their efforts have on target communities, individuals or environments. Reasons for this include: 1. Lack of organisational clarity about the differences between M&E and impact. 2. Development organisations often work through partners. They then struggle to understand the scope of their influence and the levels at which they can realistically assess impact (and what their partners should be assessing). 3. Attributing evidence of change to specific interventions is challenging, if not impossible. 4. The design of the impact assessments is too complex; or it attempts to address too many needs. 5. Identifying useful starting points from which to assess impact, including baseline data and deciding which indicators to work with. 6. There are so many tools and processes available that designers of impact assessments overly complicate the process; and/or require staff and partners to work with tools that they are unfamiliar with. 7. Honest impact assessments are hard to find. Its very hard to tell the truth if it will negatively affect chances of funding, or if it threatens relationships in any way. 8. Impact assessment findings are not used creatively or effectively, so their value is not always recognised Impact assessment 10/11th September 2013 Handouts Page 3 Handout 3: Relationship with Monitoring and Evaluation Monitoring and evaluation focus directly on the progress and effectiveness of projects and programmes, while impact assessment focuses on change. They are different but complementary processes. Each has a different purpose in relation to our work, and all are necessary. This is summarized in the chart below: Monitoring Evaluation Impact Assessment Measures on-going activities Measures performance against objectives Assesses change in people’s lives: positive or negative, intended or not Main work during implementation Main work in middle or at end of project/programme cycle Can be included at all stages and/or can be used specifically after the end of programme/project Focus on interventions Focus on interventions Focus on affected populations Focus on outputs Focus on outcomes/impact Focus on impact and change What has happened? Did we achieve what we set out to achieve? What has changed? For whom? How significant is it for them? Will it last? What, if anything, did our programme contribute? What is being done? Impact assessment 10/11th September 2013 Handouts Page 4 Handout 4: Theories of Change and Dimensions (or Domains) of Change: WHAT IS A THEORY OF CHANGE? - - - A coherent connection between a system’s mission, strategies and actual outcomes, and articulated links between those who are being served, the strategies or activities that are being implemented, and the planned outcomes A clear and testable hypothesis about how change will occur that not only allows implementers to be accountable for results, but also makes these results more credible because they were predicted to occur in a certain way An agreement among stakeholders about what defines success and what it takes to get there A framework for impact assessment with dimensions of change and a menu for indicators to explore A powerful communication tool to capture the complexity of the project initiative Note: Theories of Change can be set at Organisational levels or Programme levels. For YCI we plan to set them at organisational level WHAT’S INCLUDED IN A THEORY OF CHANGE? Four clearly articulated components: - Component 1: A conceptual piece considering how change happens in relation to target groups and issues that relate to your organisational mission and mandate - Component 2: An Organisational Change Pathway which includes an articulation of: The problem(s) to be addressed; and their underlying causes A vision of change you (your org/project) wants to effect Populations: who you are serving. Principles of Engagement Strategies: what strategies you believe will accomplish desired changes Outcomes: what you intend to achieve which will lead to these changes Contribution to your organisational goal And the relationship between all of these core elements - Component 3: An Impact Assessment Framework based on the components above – often framed as “Dimensions of Change” - Component 4: A process of critical reflection to test assumptions that were made and revise the original theory Note: Different organisations have opted to complete some or all of these components in relation to their Theories of Change DIFFERENCES BETWEEN TOC AND OTHER LOGIC MODELS: - It shows a causal pathway by specifying what is needed for goals to be achieved Impact assessment 10/11th September 2013 Handouts Page 5 - - It requires the articulation of underlying assumptions which can be tested and measured. It changes the way of thinking about implementation strategies: the focus is not so much on what is being done, but rather what needs to be changed. The logical framework should be guided by and developed as a result of the Theory of Change ( they are not mutually exclusive) Logic Models Theories of Change Representation List of Components Descriptive Critical Thinking Pathway of Change Explanatory WHY DEVELOP A TOC? - For more effective planning To improve and assure accountability To be more targeted in resource allocation To communicate and market development strategies To make direct links to action IT CAN BE USED: - As a framework to check progress towards change (to complement project logic) and to stay on course To be able to test the weak links ion the change pathway (right people? right strategies? Right outcomes?) To document lessons learned about what really changes in relation to our efforts To keep the process of implementation, and impact assessment transparent, so everyone knows what changing and how As a basis for reports to funders, policymakers, boards WHAT DO THEY LOOK LIKE - Usually they are 2-5 pages in length with a short narrative followed by a diagram. Some look like flow charts Some diagrams look like flow charts Some are much more elaborate What they must so is illustrate your organizational (or programme) pathway to change HOW TO DEVELOP A THEORY OF CHANGE - Developing a Theory of Change must involve all key stakeholders The idea and the process needs buy-in from Senior Management Usually it is wise to hire a facilitator to help you (the organization or programme) through the process of designing a process that suits your needs. With the facilitator, you will decide what research needs carrying out; and what needs to be agreed; and what needs to be so Impact assessment 10/11th September 2013 Handouts Page 6 - You will carry out the research and relevant tasks, and then probably call the facilitator back in to complete and to be assured that all parts are in order and that you have an effective plan in place for ensuring that the ToC will be useful to and used in the future. Impact assessment 10/11th September 2013 Handouts Page 7 Draft Domains of Change Framework for Education (BEP) Outcomes Outcomes Children participate in decisions that affect their education Policy and regulation ensures quality education that is affordable and accessible for all children Parents and guardians support the education of all children, especially girls and marginalised children Communities, including children are involved in supporting and influencing school policy and development Community and civil society organisations are involved in operating and managing schools Outcomes 6 Changes in ways local communities influence educational policy and contribute to the provision of education Improved enrolment rates, especially for girls and marginalised children Improved attendance and completion rates especially for girls and marginalized children All children and young people acquire socially and economically relevant skills as a result of their educational opportunities Impact assessment 10/11th September 2013 5 Changes in enrolment and completion of education for all children, especially girls and marginalised children Outcomes Adequate resources are allocated for the provision of quality education for all Budgets are used effectively for the implementation of policies (incl. teacher pay, text books, etc) 2 Changes in the provision of quality education by Governments and other mandated bodies Schools are accessible and available to all children All children feel safe on the way to school and within school environments Schools are effectively managed Facilities and resources offered within the school support effective learning 1 Changes is the ability of all children and young people to access quality education which results in relevant and useful learning for them 3 Changes in the relevance and availability of educational infrastructure and institutional development Outcomes School and educational curricula are relevant and appropriate Student assessment systems and processes are appropriate and equitable 4 Changes in the quality and relevance of educational curricula, teaching standards and assessment methodologies Handouts Page 8 Teaching force delivers the curriculum effectively Teachers’ pay and conditions enable them to carry out their duties Handout 5: Spheres of Influence Organisations working through national or local partner organisations can’t attribute their efforts to changes at community level So they struggle to understand • what they should be assessing in terms of impact • what their partners should be assessing • the relationship between these different areas. One way to address this challenge is to work out what changes can realistically be linked to your organisational efforts: consider and agree what areas of changes might link directly to your efforts; and what areas of change may be indirectly linked to your efforts. This can be reflected in your Dimensions of Change Example from CDKN: CDKN works primarily to influence changes in the outer boxes but contributes to changes in the outer and inner rings Slide 15 Impact assessment 10/11th September 2013 Handouts Page 9 Handout 6: Handout Different approaches to Impact Assessment Three broad, and sometimes overlapping, approaches are often used to assess impact: The first approach is likely to include be a retrospective study of a project which has, typically, been planned using a linear approach to change (as developed through logical framework analysis for example). More often than not, this assessment involves the use of an external (or partly external) team who carries out the assessment at a fixed point after the completing of the project. The main purpose of this approach is to test or verify whether the logic of the project) was correct: did it achieve the changes that it set out to achieve? The second approach is more process driven - and less judgmental - than the first. Stakeholders are involved in all stages of the project or programme cycle: in the design and the development of the project or programme, they are included in identifying the changes that they would like to see in their lives; during the implementation phase, they influence the direction the project takes. They are included in the monitoring and evaluation, at the impact assessment stage, they inform the process identifying the changes that have taken place in their lives as a result of the project or programme. This type of assessment is clearly empowering for stakeholders; and it makes valuable contributions to organisational learning. The third approach is somewhat different from the first two. It is a study that takes place some years after the completion of the project or programme and it sets out firstly to identify changes that have taken place within the community, and secondly, to what extent they can be attributed to particular interventions. The purpose of this type of assessment is to understand to what extent organisational efforts are making a difference to the lives of the people they claim to be working with and for. Although this type of assessment is more complex and time consuming, it is one that all organisations need to consider carrying out from time to time. All three approaches – and variations on them - have their own validity in relation to the main purpose of assessing impact. Often, a combination of elements of all three approaches will be necessary in order to understand both planned and unplanned changes. Different Strategies for Assessing Impact: The approach to impact assessment is very connected to the purpose for carrying it out. Impact Assessment is best carried out if a number of strategies are used. For example, some or all of the following could be used to assess the impact of a programme: a. Build ongoing monitoring of impact into M&E reporting formats. Impact data will then be gathered routinely alongside other M&E data. b. At certain points of the programme, build in time for critical reflection (“so what has changed so far in relation to our efforts? For whom? How significant is this? How should we adapt our programme as a result?”). This can be done at any time or specifically as part of the mid-term review and evaluation process c. Carry out one or more tracking study through the life of a programme. This will provide information about how identified sample groups are changing and developing as a result of on-going programme efforts. d. Ensure that a retrospective study takes place sometime after the completion of the programme. This will assess the changes that have occurred in relation to the programme, and to what extent it was able to contribute to these changes. Key tips: Impact assessment 10/11th September 2013 Handouts Page 10 The assessment of impact can be carried out relatively easily and cheaply. If it is done well, it can be really rewarding and motivating for all concerned. Two key tips: Ensure that they are useful, user friendly and that they build on existing structures and systems Ensure that conditions for assessing impact enable people to be involved and be as honest as possible. This means investing in time and building the trust of those with whom you are working. If people believe that you are trying to understand change in order to improve (rather than to judge success or failure), they are more likely to give honest answers. Impact assessment 10/11th September 2013 Handouts Page 11 Handout 7: Case Study Task You will be working in small groups on designing an impact assessment process for a specific thematic area or set of interventions. You will be asked to complete this task in stages through the course of the day. You will present a summary of your answers to the whole group tomorrow morning. Please document your answers either on flip chart or on PowerPoint. Note: Your aim should be to develop a process that is: Simple and user- friendly Builds on existing structures and systems Is useful for your organisational learning Before you start: Agree on case study framework: • • • • • Brief profile of org ( size, mandate, scope of activity and areas of work) Direct Interventions or working through partners? Reason for conducting an impact assessment Programme Goal Three outcomes Case Study Task 1: Designing the impact assessment What will be the main purpose of doing this impact assessment? Organisational learning? To meet donor demands? Accountability to stakeholders? For advocacy Which approach (or combination of approaches?) to Impact Assessment would be most appropriate in this Why? What will be the scope and scale of this assessment? Case Study Task 2: Develop/confirm theory of change and/or dimensions of change and areas of enquiry What is your realistic “scope of influence”? Which what areas of impact will you realistically be able to “assess”, and which areas of change will you be able to “illustrate contributions to change”? Based on this,what “Dimensions of Change” will you be looking to assess? Develop a menu of areas of enquiry which will enable you to set baselines and track progress in relation to impact Case Study Task 3: Select methods for gathering relevant information Propose a range of appropriate methods that you could use to gather relevant data for both monitoring and assessing impact (including building on or adapting existing tools and mechanisms). Impact assessment 10/11th September 2013 Handouts Page 12 Handout 8: Developing Areas of Enquiry to explore Medium and Long Term Change Areas of enquiry are written in neutral language to allow for the reporting of unintended and unexpected changes as well as those that were intended and positive.They are directly linked to Dimensions of Change and specific medium and long term changes that the project/programme hopes to achieve. Areas of enquiry are likely to vary across regions, nations and sectors. It is essential that the indicators are relevant to specific situations E.g. using changes in awareness and mobilisation of key stakeholders (in relation to child labour). Dimension: Changes in frequency and content of CL issues discussed in public arenas and on political agendas Dimension Changes in awareness and attitudes of key stakeholders Dimension Changes in the extent and type of Community involvement around CL issues Possible Areas of Enquiry The extent to which leaders and other respected people are seen to be championing CL The ways in which the media – TV, radio, electronic (blogs), newspapers etc – covers CL issues Numbers and types of civil society campaigns on CL issues How and when CL appears in other public forums The extent to which CL issues are included in school curricula The extent to which universities take an interest in CL Possible Areas of Enquiry Shift in awareness, knowledge, attitudes and commitment in respect of CL among the general public and key multiplier groups in particular e.g. influencers and decisionmakers. Tools/ Sources of Information Interview/group discussions s with stakeholders and documentary review. Possibly network analysis. Media monitoring over an extended period Interviews/group discussions with stakeholders and documentary review Public record Interviews with educators Sources of information KAP-type surveys preferably linked to baseline information relating to changes. If no pre-survey, sample of depth interviews/group discussions to probe further the extent of change and the stimuli for the changes. Sources of information Interviews/group discussions with stakeholders Possible Areas of Enquiry Shifts in levels of Community mobilisation around CL - e.g. CL committees set up and operational Levels and types of actions taken by Participative enquiry and communities taking effective action e.g. self- interviews. sensitizing, monitoring, reporting, self enforcing/regulating where appropriate, referring vulnerable children and households to agencies Levels of mobilisation against CL in schools Ways in which children and households acquire social and psychological coping mechanisms Impact assessment 10/11th September 2013 Handouts Page 13 Handout 9: Example of a Success Scale - Advocacy Programme Impact assessment 10/11th September 2013 Handouts Page 14 Handout 10: Setting Baselines for Impact Monitoring and Assessment Why have a baseline? The purpose of impact monitoring and assessment is to involve programme leaders and managers and key stakeholders in a coherent process of understanding and recording the changes that are taking place in relation to their organisational or programme efforts. This information will enable them to: understand how and to what extent programmes are contributing to change objectives identifies in relation to organisational Dimensions of Change; communicate success and understand where (and why) particular programmes are failing to achieve their planned impact; and adapt or adjust programmes and projects in order to achieve greater positive impact. How should the baseline be developed? Baselines will be characterised by the following guiding principles. The methodology should be flexible. Baselines will be tailored for each programme depending on the strength of existing relationships, knowledge and capacity. Baselines, and subsequent impact monitoring, will be designed to capture lessons learned and incorporate these into future methodologies. Baselines will be “light-touch” in character. The emphasis will be on analysis. They will form an intelligent interpretative view of where the programme stands in relation to relevant dimensions of change, as well as the programme change objectives and indicators. Baselines should not be a heavy resource intensive or expensive exercise. Baselines should build on or draw from other diagnostic processes, (for example situation analyses which inform programme design). They should not duplicate other processes. They should be developed and written in the most useful language for the programme, but may also need to be translated into English for purposes of global reporting. What information should be covered by the baseline? The baseline will basically provide information on the current situation, so that changes at a later date can be assessed against that situation. The precise nature of the information to be collected will depend on the type of programme, whether or not it is time bound, how far desired changes can be accurately predicted, and the scale of resources dedicated to the programme. The following should be considered: All programmes should develop a degree of baseline information against relevant dimensions of change If there are specific change objectives associated with the programme then the baseline should attempt to show the current situation regarding those particular change objectives. Who should be involved in collecting baseline information? Impact assessment 10/11th September 2013 Handouts Page 15 The programme leader and programme managers should collect the information in collaboration, where possible, with key stakeholders. This can be a large or small exercise depending on the size and duration of the programme. Where there is a programme or country engagement leader, s/he would be expected to coordinate this role. However, key partners, stakeholders, and/or different target groups should also be represented wherever possible. How should information be collected? The process for data collection will vary according to need, on a case-by-case basis. It might also vary according to the type of tools and methodologies used to access the data. At its most basic, it might include a desk exercise which is informed by a few key interviews. It might also involve a dedicated workshop for key stakeholders and representation from target groups. At the other end of the scale, it could involve a more strategic and holistic planning process When should it be collected? Often baseline information proves to be a strenuous time consuming exercise which turns out to be of little use for impact monitoring. This is because the exercise was carried out too early - before the programme implementation strategy had been clearly defined and articulated. In order for it to be useful as a basis for collecting impact monitoring information, the programme needs clarity on its priority dimensions of change and its implementation strategy for each of these dimensions. It is therefore recommended that the baseline information gathering process is carried out as an integral part of the programme planning process. Under some circumstances, this may mean it is carried out some months into the programme. Impact assessment 10/11th September 2013 Handouts Page 16 Handout 11: Criteria for Selecting Tools The following criteria may be useful in assisting staff to select appropriate tools for differing impact assessment needs. It is important to stress that tools are only useful if they are used intelligently! Outcome or impact: To what extent does the tool seek to study impact, and not only outcomes? Measurement (Quantitative Data): How far does the tool allow for quantitative analysis of change? Description: How far does the tool allow for qualitative analysis of change? Attribution of change: How well does the tool deal with attribution of change and the exploration of cause-effect-relationships? Baselines: To what extent does the tool cope without the existence of baseline data? Use of Indicators; How far does the tool cope without the existence of pre-determined indicators? "Proving", giving evidence for accountability: To what extent does the tool seek to provide evidence of change? "Improving", promoting critical reflection: To what extent does the tool promote critical reflection among participants? Local participation: To what extent does the tool explicitly take a participatory and empowering approach that includes the people intended as ‘beneficiaries’? Aggregation: To what extent does the tool allow for aggregation of outcomes or impacts (across groups affected by an intervention, e.g. geographically, or at an organisational level)? Disaggregation: To what extent does the tool support differentiated analysis of change among different groups affected, e.g. socio-economic, ethnic, cultural groups? Gender disaggregation: To what extent does the tool specifically support differentiated analysis of change by men, women, boys and girls? Use by implementing staff: How appropriate is the tool for direct use by front-line implementing staff? Use by communities: How appropriate is the tool for direct use by communities? How appropriate is the tool for direct use by people with limited literacy? Transparency and feedback: To what extent does the tool incorporate feedback on findings to implementing staff and those being assessed? Sector coverage: To what extent is the tool applicable across more than one sector? Impact assessment 10/11th September 2013 Handouts Page 17 Handout 12: Some Key Data Collecting Methods and their Strengths and Weaknesses (Source: INTRAC) Method Definition and Use Case Studies Collecting information that results in a story that can be descriptive or explanatory and can serve to answer questions of how and why Strengths Weaknesses Can deal with a full variety of evidence from documents, interviews, observation Can add explanatory power when focus is on institutions, processes, programmes, decisions and events Good case studies are difficult to do Require specialised research and writing skills to be rigorous Can’t generalise findings to population Time consuming Difficult to replicate Focus Groups Interviews Holding focused discussions with members of target population who are familiar with pertinent issues before writing a set of structured questions. The purpose is to compare the beneficiaries’ perspectives with abstract concepts in the evaluation’s objectives Similar advantages to interviews Can be expensive and time consuming Particularly useful where participant interaction is desired Must be sensitive to mixing of hierarchical levels Useful way of identifying hierarchical influences Can’t make gereralisations The interviewer asks questions of one or more persons and records the respondents’ answers. Interviews may be formal or informal, face-to-face or by telephone, or closed- or open ended. People and institutions can explain their experiences in their own words and setting Time consuming Impact assessment 10/11th September 2013 Flexible to allow the interviewer to pursue unanticipated lines of enquiry or to probe Handouts Page 18 Can be expensive If not done properly, the interviewer can They can be structured, semi-structured or – rarely – unstructured influence the interviewee’s response issues in depth Particularly useful where language difficulties are anticipated Greater likelihood of getting input from senior officials Observation Questionnaires Observing and recording situation in a log or Provides descriptive information on context diary. This includes who is involved; what and observed changes happens; when, where, and how events occur. Observation can be direct (observer watches and records), or participatory (observer becomes part of the setting for a period of time). Quality and usefulness of data highly dependent on observer’s observational; and writing skills Developing a set of survey questions whose answers can be coded consistently The quality of responses highly dependent on the clarity of questions Can reach a wide sample simultaneously Allows respondents time to think before they answer Can be answered anonymously Impose uniformity by asking all respondents the same things Findings can be open to interpretation Does not easily apply within a short time-frame to process change Sometimes difficult to persuade people to complete and return questionnaire Can involve forcing institutional activities and people’s experiences into predetermined categories Make data compilation and comparison easier Written Reviewing documents such as records, administrative databases, training materials Impact assessment 10/11th September 2013 Can identify issues to investigate further and provide evidence of action, change and impact Handouts Page 19 Can be time consuming Document Analysis Stories of Change KAB/P studies Media monitoring and correspondence to support respondents’ perceptions Can be inexpensive These are similar to case studies but with a greater focus on change. They are many variations. MSC is a specific process that mobilises small groups involved with interventions in the regular capturing of outcome stories. (Davies, Dart 2005) Very good for capturing significant, unexpected , positive and or negative changes Time consuming in preparation, implementation and analysis Very participatory Not useful for collecting quantitative data A KAB/P survey is a method of obtaining largely quantitative data relating to people’s awareness, knowledge, attitudes, behaviour, practices or some other aspect of their lives. It is usually a sample survey using a structured questionnaire. It can be administered directly face-to-face or by telephone, or through self-completion via the internet or physical documents. KAB/P surveys can be used longitudinally to collect baseline data and capture changes without relying on interviewees’ recall Time consuming in preparation, implementation and analysis Media monitoring encompasses a range of processes for tracking the appearance in the media of matters of interest (e.g. child labour issues). This is typically outsourced to an agency and is increasingly employs electronic search technology. Very useful for tracking change in relation to advocacy efforts as change is notoriously unpredictable Cost implications if outsourced. Is only useful if high levels of analysis applied Participatory approaches encompass a range of methodologies which ensure that Particularly appropriate in working with groups for whom questionnaires or conventional FGDs Not so good for capturing information when number of stakeholders is very Impact assessment 10/11th September 2013 Very good for tracking and assessing changes in capacity Handouts Page 20 PRA techniques that the perspectives and insights of all stakeholders, beneficiaries as well as project implementers, are taken into consideration in the design and conduct of evaluative research Examples include risk maps, time lines, scoring and ranking exercises Impact assessment 10/11th September 2013 are less appropriate. They encourage people to describe aspects of their lives and changes brought about by interventions in their own terms. Can be used to convert qualitative data into quantitative information Handouts Page 21 high Time consuming and requires skilled facilitators Impact assessment 10/11th September 2013 Handouts Page 22 Handout 13: A word on RCTs: Definition of Randomised controlled Trial (RCT) - Medinet Randomized controlled trial: (RCT) a study in which people is allocated at random (by chance alone) to receive one of several clinical interventions. One of these interventions is the standard of comparison or control. The control may be a standard practice, a placebo ("sugar pill"), or no intervention at all. Someone who takes part in a randomized controlled trial (RCT) is called a participant or subject. RCTs seek to measure and compare the outcomes after the participants receive the interventions. Because the outcomes are measured, RCTs are quantitative studies. In sum, RCTs are quantitative, comparative, controlled experiments in which investigators study two or more interventions in a series of individuals who receive them in random order. The RCT is one of the simplest and most powerful tools in clinical research. In relation to Development In recent years there has been a lively debate in impact assessment about the circumstances in which it is possible to attribute changes to interventions – to demonstrate causality. This debate has highlighted the role of experimental – or quasi-experimental methodologies in impact assessment. These methodologies use a control group – a “counterfactual”1 - which is identified in advance and does not have any contact with the or project. What is more, in the most pure version of the methodology, programme targets – usually people - are assigned at random between the control group and the group that receives the intervention, so as to avoid selection bias. These methodologies in their pure form are usually referred to as Randomised Control Trials or RCTs. Experimental methodologies work best in simple, short-term, single-variable, interventions. In these contexts it may possible to attribute changes to the intervention. Interventions typical of the enabling environment (e.g. capacity building, governance, participation, and advocacy) are usually complex, long-term and multi-variable. These are usually not susceptible to experimental methodologies such as RCTs because the links between the interventions and the changes cannot be proven statistically. However it is possible to demonstrate a contribution or credible association 2 between the interventions and the changes that the impact assessment identifies using other methodologies. These include • Identifying the intervention logic. Capturing data from a sufficient number of different sources, both quantitatively and qualitatively, to enable “triangulation”. Triangulation reduces the chance of bias in any particular data source. Establishing plausible connections between the interventions and the changes identified. Exploring other explanations for the changes with an open mind before eliminating them. A “counterfactual” is a situation - e.g. the status or condition of a group of people - that exists independent of the programme. Control groups are the most commonly used counterfactuals in impact assessment and evaluation 1 Impact assessment 10/11th September 2013 Handouts Page 23 The most important way of shining a light on the connection between the interventions and the changes that are observed is a systematic impact assessment design process, and in particular the establishment of the underlying intervention logic or Theory of Change Harry Jones, ODI, comments on some of the problems associated with RCTs: RCTs tend to be carried out where they are methodologically convenient rather than where the new knowledge is really needed. There is massive publication bias, by our count over 95% of published RCTs show ‘positive’ impact, which severely limits the ability to really learn The key problem is about the relationships between RCTs and policy. RCTs were hugely fashionable in the US in the ’60s until it was clear that they couldn’t deliver all that was hoped. Now, it seems that development policy makers are in love with the idea of having clear numbers on the impact of all of their work. The problems include: • • • • Policy makers often assume a much higher level of external validity than is actually appropriate, and tend to ignore the careful caveats which come along with RCTs The RCT model is only suitable for measuring impact in a subset, nay a minority of the kinds of intervention required for development. However they are being given a disproportionate amount of attention There is an issue of cost-effectiveness - about where we choose to spend evaluation budgets and with what coverage There is a risk that by demanding ‘rigorous impact evaluation’, organisations feel compelled to use methodologies like RCT. This may suit output-driven programmes like distributing bed-nets, but, as stated above is inappropriate for interventions such as governance, capacity building, budget support, policy influencing.” Impact assessment 10/11th September 2013 Handouts Page 24 Handout 14: Guidelines on Understanding and Reporting Changes (by Nigel Simister) Almost all planning, monitoring and evaluation (PME) systems require some element of formal reporting. Regular reports need to be written both for external stakeholders, such as donors, and for internal learning. One of the most essential elements of many reports is the description of changes that are occurring within a project or programme. However, many organisations are weak in this area. Too often, reports that are intended to deal with changes are inadequate. Specifically: reports focus entirely on outputs or activities, not changes; they focus only on positive changes, not on negative changes or areas where expected change has not happened; they report change in an anecdotal way, with insufficient evidence to back the changes up; or reports claim that an organisation is responsible for changes that are more due to external influences. Sometimes this can be due to inadequate PME systems which fail to identify change. However, often it is simply an issue of how change is reported. When working in a project or programme it can be difficult to assess how much information to provide. Too much, and the reporting of the change is buried under pages and pages of reports. Too little, and it is hard for someone without an in-depth knowledge of the project or programme to understand the importance or relevance of the change. When producing brief reports for an external audience (i.e. for people who are not working within your project or programme, and perhaps not even within your own region), you should be looking to provide enough information about individual changes for other people to make a considered opinion about the changes resulting from your work, and about any lessons learned. The list below provides some of the areas that should be considered when reporting on a change or changes. Has there been change? The basic question to address is whether or not things have changed. The changes could be positive or could be negative. In some cases it may also be worth specifically noting changes that have not happened but which should have been expected. How significant was the change? Some changes are relatively minor, whilst others are major and life-changing. It is worth considering whether you need to emphasise the significance of the change within your report. How many organisations/people were affected by the change? Sometimes changes are reported across a number of people or organisations. At other times, you might be reporting a case study based on just one or two people/organisations. In either circumstance, it is useful to know roughly how many people/organisations you think might have been (or will be in the future) affected by the change. Which target groups were affected by the change? Change does not normally happen equally across all types of stakeholders. Some may benefit more than others. A report should be clear about which particular target groups were involved in the change. What was the impact on sub-categories or groups? A report should emphasise any differences between different target groups, if known. For example, some Impact assessment 10/11th September 2013 Handouts Page 25 changes might affect only specific communities. Simply reporting on change across a large number of different groups might hide significant differences. Was the change intended or not? This can be an extremely valuable source of learning. Sometimes the most profound changes are those that were unplanned. Describing changes of this kind can provide valuable lessons to feed back into planning cycles. Is it likely to be sustainable? Some changes might be long-lasting, whilst others might be relatively short-lived. In some circumstances it might be useful to provide an estimate of how sustainable any reported change is likely to be and to indicate any risks or assumptions that might influence the sustainability. What made the change happen? Reporting on change by itself is interesting but rarely useful for learning purposes. However, if there is an assessment of how the change came about, or what were the key processes leading to it, others may be able to replicate the work (or avoid mistakes in the case of negative changes). Reporting on the key processes that led to a change will also help to substantiate any claim that the change was a result of a particular project or programme’s work. How might the change result in further changes in policies/organisations/people’s lives? Sometimes the implication of change is unclear to outsiders. If you report on an impact – a significant or sustainable change – in people’s lives then it is usually clear what the benefits are. However, if you are reporting on outcomes – the immediate changes resulting from your work – the significance of change may not be clear to everyone. For example, you might report that a government department has higher capacity; a new policy on adaptation has been developed; or there has been increased collaboration between different stakeholders. Within your project/programme the implications may be clear. But for an outsider you might need to spell out why you consider this an important change, and what you hope the ultimate long-term result (or impact) will be. This is basically the “so what?” question. How do changes compare to the baseline (if any)? If you report that 75% of people in a location now have access to clean water, this could be considered as an extremely important change. On the other hand, the situation might be worse than last year! Wherever possible, a report describing change should detail the original situation so that people can understand how large or important the changes are. This applies to both quantitative and qualitative changes. How do changes compare to what was hoped for, or considered realistic? Equally, if you report that 15 national policies were influenced by your efforts the implication is that this is a positive change. However, if you planned to influence 30 policy initiatives, this casts a different light on the information. It is therefore often useful to describe what was originally planned for, so that people reading your report can see immediately the scale of any change relative to your expectations. What evidence do you have for your change? This is arguable the most important aspect to report when describing any change or changes. There is a world of difference between describing the findings of a professionally-conducted, large-scale research study, and reporting findings based on a conversation with a couple of people. The description of evidence does not have to be substantial. It is enough to make an introductory statement such as “the findings of focus-group studies with government officials suggested that …” or “anecdotal evidence suggests that …” or “independent research by government bodies has found that …” This will allow the reader to make up his/her own mind about the value of your evidence. N.B. There is no reason at all why anecdotal evidence of change should not be described in a report. Provided it is clear that the change reported is not based on rigorous data collection and analysis methodologies, impressions of change Impact assessment 10/11th September 2013 Handouts Page 26 can still be useful. The danger comes when anecdotal evidence is reported as if it were a firm conclusion based on rigorous evidence, instead of tentative conclusions which needs to be further investigated if it is to be properly validated.) How was any change attributable to the work of your organisation? Unless reporting on the external socio-economic environment, you will probably have reason to believe that at least some of the reported changes are attributable to your organisational efforts, or the work of your partners. It is therefore useful to describe how you think your project or programme contributed to any change. Where necessary, you can also describe other factors or organisations that may have contributed. With what degree of confidence can you state the change? PME systems often encourage people to be very definite in their opinions. For example, a logical framework encourages people to say whether a change has happened or not. In many cases, however, you may have some evidence that a change has occurred, but you may not be sure, or you might be sure the change has occurred, but not sure how far your organisation has contributed towards it. In these cases, it is usually better to state the change anyway, and to add some qualifying statements that make it clear how confident you are that change has occurred. If you think there are other possible explanations for why change has happened, it is often useful to state this as well. Again, anyone reading your report can make up his/her own mind provided they have the necessary information on which to base an opinion. Of course it will not always be necessary to report on each of these areas for every single change. Otherwise your reports will be hundreds of pages long, and nobody will ever read them! However, you should seek to ensure that you at least provide enough information so that people can make up their own minds about the value or importance of the changes you describe, or provide references to source material so that people can investigate further if they so choose. Impact assessment 10/11th September 2013 Handouts Page 27 Handout 15: How to build Impact Monitoring into Existing Systems and Processes The purpose of this exercise is for you to consider to what extent your current planning, monitoring and evaluation systems and processes factor in assessment of change; and how they would need to be developed in order to do so effectively. 1. Based on the series of questions below, please identify: What elements are already in place in your organisation (and in what form – guidance? Specific mention in forms? Policy docs??) Prioritise what you need in order to be able to monitor impact effectively in your projects/programmes 2. Consider how you might be able to meet these needs in the near future Planning Do you clearly articulate the changes (at outcome and impact levels) that you aim to achieve in your projects/programmes? Do you design your M&E and impact assessment strategies in the planning stages of your project/programme? Which stakeholders, if any, do you include these processes (e.g. other members of staff, partners, community representative, others?) Developing Baselines: Do you currently conduct any baselines? If so, do they include a focus on medium and long term changes that the programme expects to achieve? If so, do you identify areas of enquiry and methods for gathering information relevant for monitoring impact? Do you compile and analyse this information? Monitoring: Do your monitoring processes systematically explore changes relating to the project/ programme? If so, do you analyse this information and make choices about future direction based on this analysis? Which stakeholders, if any, do you include these processes (e.g. other members of staff, partners, community representative, others?) Reporting: How do you use monitoring data in your reports? Do you clearly evidence and analyse changes that have been noted? Do you include aspects of learning in these reports? Are you able to effectively communicate the decisions that you are taking as a result? Capacity, Time and Resources: Do staff members have the required skills to design and implement these impact monitoring processes? Are resources and/or time a constraint? Impact assessment 10/11th September 2013 Handouts Page 28 Handout 16: Checklist to ensure Organisational Learning from Evaluations and Impact Assessments Developing Organisational Memory The organisation has mechanisms for remembering the experience of its current and previous work through the development of highly accessible databases and resource information centres All written reports and key documents are cross referenced and made easily accessible to all staff The organisation is not vulnerable to losing information when staff leave The organisation has a systematic database of all its project and programme work which enables staff and outsiders to identify where expertise can be found within the organisation The information function is given prominence and is resources adequately to enable the organisation to keep its records up to date Applying the Learning The organisation systematically uses its learning to improve its own practice and influence the policy and practice of other organisations or agencies The organisation writes up and publishes its experience for a wider readership without using unnecessary technical jargon The organisation has a strategy for scaling up its impact which reflects the learning it has developed on “what works” The organisation changes its practice and priorities to reflect now knowledge and insights, in an effort to constantly improve its effectiveness The organisation is constantly building its capacity and innovating based on what it has learned Question: How does your organisation measure up? Impact assessment 10/11th September 2013 Handouts Page 29 Handout 17: Guidance on writing a TOR for Contracting a Consultant to carry Out an Impact Assessment (NB There may be law in your country and/or organisational regulations and procedures governing these matters.) Pre-amble This should mention the commissioning agent(s) and summarise the important about the impact assessment in one paragraph. Context: This should contain an introduction to the programme Theory of Change in relation to it. If there is no theory of change, the context should include the elements contained in it: problems the programme seeks to address, underlying causes of those problems, programme vision and principles of engagement, who it works with, and how to achieve which medium and long term change; and the overall goal of the programme If there is a logical framework, this should also be included here. Purpose There should be a statement about the purpose of the impact assessment and the proposed audiences and uses. Scope, main focuses, and scale of the impact assessment This section addresses – in varying amounts of detail and depth - what the assignment will assess. This will include the Dimension of Change, and possibly the Areas of Enquiry and key assessment questions. However it is important to emphasise that these will be finalized through dialogue with the impact assessment team in the inception phase. Information sources and methodologies This section should develop a provisional list of known data sources for the impact assessment and propose some key methodologies for gathering relevant information. Again, it is important to leave the door open for the assessment team to propose other tools and methods. References to any relevant baseline for impact data should be included here. There should be a statement about expected stakeholder participation. Impact Assessment Process and deliverables, and their timings This section includes the proposed phases of the impact assessment – principally inception, data gathering, analysis, report writing and presentation - the nature of the deliverables. Accurate timings of the phases and the deliverables should be stated. Supervision and responsibilities Impact assessment 10/11th September 2013 Handouts Page 30 This section includes the management arrangements for the assessment and the respective responsibilities of the commissioners and the assessment team. Required skills, experience, capacity This section sets out expectations of the size and nature of the impact assessment team, and the respective skills and experience of the members. It can set out the parameters of sub-contracting on the part of the assessment team. Budget range An indicative budget is an important piece of information for applicants. This section can also indicate what costs will be met and payment arrangements and timings. How to apply This section should specify what is required of applicants in their submissions. Dates and other arrangements should be included. Information about short-listing and selection is also important. Impact assessment 10/11th September 2013 Handouts Page 31 Handout 18: Example of a work plan structure An impact assessment work plan will usually contain the following sections. 1. Overview of the interventions including the logic model or Theory Of Change 2. Purpose, audience and uses of the evaluation 3. The assessment framework The high-level focuses of the assessment (the Dimensions of Change). The more detailed areas of enquiry to support investigation against Dimensions of Change The data sources and the methods of collection - including baseline information. 4. Information Collection and Analysis Types of instrument, including for example the key interview guides Approaches to sampling Lists of key informants; and target groups for quantitative surveys 5. Roles and responsibilities A detailed matrix 6. Assumptions, risks and challenges for the assessment Strategies for risks and challenges 7. Work scheduling Refer to dependencies. 8. Deliverables Structure of the report(s) and other deliverables Process for quality assurance. Impact assessment 10/11th September 2013 Handouts Page 32 Handout 19: Further Reading and Web Sites for Impact Assessment Annie E. Case Foundation, A Handbook of Data Collection Tools for Advocacy and Policy ACT Organisational Capacity Assessment Tool, February 2008 ACT: A guide to assessing our contribution to change O’Flynn, Maureen: Impact Assessment: Understanding and Assessing our Contributions to Change, INTRAC 2010 O’Flynn, Maureen: Tracking Progress in Advocacy: Why and How to Monitor and Evaluate Advocacy Projects and Programmes , INTRAC 2009 Simister Nigel: Monitoring and Evaluation of Capacity Building –Is it Really that Difficult? INTRAC 2010 Bamberger, Michael. Conducting quality impact evaluations under budget, time and data constraints. World Bank, Washington DC, 2006. CAFOD et al. monitoring government policies: A toolkit for civil society organisations in Africa. London (2006?) http://cdg.lathyrus.co.uk/docs/MonitorGovPol.pdf Chambers, Robert et al. Designing impact evaluations: different perspectives. International Initiative for Impact Evaluation, Delhi, 2009. Chapman, Jennifer and Mancini, Antonella. Impact assessment, drivers, dilemmas and deliberations. Sight savers International, Hayward’s Heath, 2008. Chapman, Jennifer and Wameyo, Amboka. Monitoring and evaluating advocacy: a scoping study. ActionAid, London, 2001. Davies, Rick and Dart, Jess. The ‘Most Significant Change’ (MSC) Technique: A Guide to Its Use", 2005. http://www.mande.co.uk/docs/MSCGuide.htm DFID. Monitoring and evaluating information and communication for development programmes. London, 2005. European Evaluation Society: The importance of a methodologically diverse approach to impact evaluation. EES. 2007. European Union. Evaluating socio-economic development. Sourcebook 2. Methods and approaches; participatory approaches and methods. Brussels. 2003. Garbarino, Sabine and Holland, Jeremy. Quantitative and qualitative methods in impact evaluation and measuring results. Governance and Social Development Resource Centre. Birmingham 2009. Grant, Jonathan et al. Capturing Research Impacts. A review of international practice. Rand Corporation, Santa Monica, 2010. http://www.rand.org/content/dam/rand/pubs/documented_briefings/2010/RAND_DB57 8.pdf. Jones, Harry. A GUIDE to monitoring and evaluating POLICY INFLUENCE. ODI, London, 2011. Jones, Harry and Hearn, Simon. Outcome Mapping: A realistic alternative for planning, monitoring and evaluation, ODI, London, 2009. http://www.odi.org.uk/resources/details.asp?id=4118&title=outcome-mapping-realisticplanning-monitoring-evaluation Jones, Nicola et al. Improving impact evaluation production and use. ODI. London 2009. Impact assessment 10/11th September 2013 Handouts Page 33