Stockholm October 18, Johan Harvard
Workshop:
Evaluations as a policy instrument for sustainable development
How can we improve the learning process in political decision-making?
13.00 Welcome
Johan Harvard, Swedish Agency for Growth Policy Analysis
13.00
– 13.20
13.20 – 13.40
The Power of Knowing: Why Improved Evaluation and Learning are
Needed to Realise the Green Future
Johan Harvard, Swedish Agency for Growth Policy Analysis
Dealing with Complex Policy Evaluations – The European Social
Fund
Steffen Ovdahl, Ramböll Management Consulting
13.40 – 14.30
Plenary Panel Discussion
Comments by
Erik Fahlbeck, Chief Analyst, Economic Analysis Secretariat, Ministry of Enterprise
Barbro Widerstedt, Analyst, Swedish Agency for Growth Policy Analysis
Steffen Ovdahl, Senior Consultant, Ramböll Management Consulting
Stockholm October 18, Johan Harvard
Presentation:
The Power of Knowing: Why
Improved Evaluation and
Learning are Needed to
Realise the Green Future
• Lots of policy movement in the SD/GG-field, in developed countries
• Complex challenges – complex policy
• Policy is not sufficiently/appropriately evaluated
• Lessons not learned – inefficient policy and wasted tax money
• However – SD/GG policy very complex to evaluate
• Thereby – SD/GG policy evalutions difficult to use for policymakers
• Evaluation vs Monitoring vs Benchmarking
• Evaluating process/implementation, results, impact
• Factual vs Counterfactual : Need to assess additionality
• European Union’s Growth Strategy for 2020
• Korea’s National Strategy and Five-Year-Plan for green growth
• Green development focus of China’s Five-Year-Plan
• South Africa’s New Growth Path and Green Economy
Accord
Status
1
Needs highlighted
• Wish to stimulate sustainable growth: push for action
2
3
4
5
Analysis preceding policy action
• Ex-ante analysis
Policy Action
• Decisions are made
Implementation
• Execution of policy
Learning after/during implementation
• Lessons learned
New policy cycle
6
• Local/regional level
– Yes : Some monitoring/evaluation of implementation and results
– No : Impact evaluations are very rare
• National level
– Yes: Benchmarking, monitoring macro indicators
– No: Impact evaluations, system evaluations linking policy to outcomes – very rare
• If you do not measure results, you cannot tell success from failure.
• If you cannot see success, you cannot reward it.
• If you cannot reward success, you are probably rewarding failure.
• If you cannot see success, you cannot learn from it.
• If you cannot recognize failure, you cannot correct it.
• If you can demonstrate results, you can win public support
Source: Adapted from Osborne & Gaebler 1992.
• SD/GG policy is inherently complex
– Policy on all levels simultaneously (local/regional/national/ international)
– Very different types of policies implemented simultaneously
(supply/demand, skills/capital/incentives/market)
– External factors very important (economic turmoil, global market demand, competition)
– Conflicting objectives (economic growth vs sustainability)
– Multiple policy areas – many different actors in play
– Interaction effects
• Simple problems Like baking a cake. A few basic techniques but once mastered likelihood of success is high
• Complicated problems Like sending a rocket to the moon. No straightforward recipe often require multiple people and teams.
Unanticipated difficulties are frequent. Timing and coordination become serious concerns
• Complex problems Like raising a child. Once you learn how to send a rocket to the moon. You can repeat and improve the process.
One rocket is like another rocket. But raising one child is not like raising another. Each is unique. Once child may require a completely difference approach to the previous one. This brings up another feature of highly complex problems. Their outcomes remain highly uncertain.
Source: Rogers (2008)
• Qualitative evaluation
– Contribution analysis
– Theory-based approach
– etc
• Quantitative evaluation
– Comparison groups (deal with selection bias)
– Quasi-experimental methods (PSM, Discontinuity etc)
– Difference-in-difference (if unobservables are time invariant)
• There is no one solution need combinations – mixed methods
• Currently – partial evaluations are carried out – leading to inability to draw clear conclusions
Difficult to use these evaluations
Patton:
• ”Evaluations should be judged by their utility and actual use; therefore, evaluators should facilitate the evaluation process and design any evaluation with careful consideration of how everthing is done, will affect use ”
• Focus on intended use by intended users
• Mutual responsibility of users and evaluators
• Encompassing 4x100 Million SEK over 2011-2014, currently 18 programs launched to help business grow
– Include supply and demand side policies: export advisory services, innovative public procurement, cleantech incubation, inward FDI, commercialization & export of public sector expertise, establishing businessresearch development consortia’s, matching of investorsinvestees, etc.
• Evaluation:
– Process, Results, Impacts
– Top-down + Bottom-up
– Quantitative + Qualitative
– Utilization: Hoping for process use as well as instrumental use
• Need of systematic improvement of evaluations – more impact evaluations
• Authorities should not underestimate the complexity and importance of impact evaluations
– Need better incentives? Will cost more initially!
• Evaluators should improve their use of best methods for impact evaluation – including use of quantitative methods
• We need more and better evaluations on all levels – how can this be achieved?
– Incentives for impact evaluations?
– How can we make sure evaluations include economic and environmental aspects – not either or?
• How can the procurement of evaluations improve, to facilitate better evaluations and improve evaluation use?
• Complex evaluations - difficult to use for policymakers: how can this be overcome? Capacity for use?
• Could/should policy be designed differently, to facilitate evaluation?
– …if a changed design would compromise quality?