What is evaluation? - Department of Industry, Innovation and Science

advertisement
EVALUATION STRATEGY
2015-2019
INDUSTRY.GOV.AU
Further information
For information on other department initiatives please see the department’s website at:
www.industry.gov.au/OCE
For more information or to comment on this publication please contact:
Martine Rodgers
Evaluation Unit
Department of Industry and Science
GPO Box 9839
CANBERRA ACT 2601
Telephone: +61 2 6213 7194
The views expressed in this publication are those of the authors and do not necessarily reflect those
of the Australian Government or the Department of Industry and Science.
© Commonwealth of Australia 2015
ISBN: 978-1-925092-67-7 (print)
ISBN: 978-1-925092-68-4 (online)
This work is copyright. Apart from use under Copyright Act 1968, no part may be reproduced or
altered by any process without prior written permission from the Australian Government. Requests
and inquiries concerning reproduction and rights should be addressed to
chiefeconomist@industry.gov.au. For more information on Office of the Chief Economist research
papers please access the department’s website at: www.industry.gov.au/OCE.
Creative Commons Licence
With the exception of the Coat of Arms, this publication is licensed under a Creative Commons
Attribution 3.0 Australia Licence.
Creative Commons Attribution 3.0 Australia Licence is a standard form license agreement that allows
you to copy, distribute, transmit and adapt this publication provided that you attribute the work. A
summary of the licence terms is available from http://creativecommons.org/licenses/by/3.0/au/deed.en.
The full licence terms are available from http://creativecommons.org/licenses/by/3.0/au/legalcode.
The Commonwealth’s preference is that you attribute this publication (and any material sourced from
it) using the following wording:
Source: Licensed from the Commonwealth of Australia under a Creative Commons Attribution 3.0
Australia Licence. The Commonwealth of Australia does not necessarily endorse the content
of this publication.
Contents
Introduction
1
Performance measurement and reporting
2
Impact on evaluation activity
3
What is evaluation?
4
Good evaluation practices – a consistent approach to evaluation
5
Governance
6
Executive Board
6
Steering Committees and Reference Groups
6
Evaluation and Audit—what is the difference?
6
Assurance and Audit Committee
6
Internal Audit role
7
Evaluation Unit role
7
Planning for evaluation across the policy and programme lifecycle
8
Evaluation types
8
Evaluation readiness
9
Prioritising evaluation effort
10
A four-year evaluation plan
10
Alterations to the evaluation plan
10
Scaling evaluation
11
Responsibility for conducting evaluations
11
Evaluation approach
12
Lessons learned — taking advantage of completed evaluations
13
Building capacity and capability
14
Fostering a culture of evaluative thinking
14
Building evaluation capability
14
Supporting guidance material
15
Measures of success
16
Evaluation maturity
16
Reviewing the Evaluation Strategy
18
APPENDIX A – Glossary
APPENDIX B – Departmental guidance materials
19
Error! Bookmark not defined.
Introduction
This Evaluation Strategy provides a framework to guide the consistent, robust and transparent
evaluation and performance measurement of programmes and policies in the department.
Evaluations, reviews and performance monitoring provide assurance that policies and programmes
are delivering outcomes as intended, performance is tracked—allowing for correction to occur—and
inform future policy and programme design. As Australia is called to adapt to changing economic and
policy environments, the evidence gained from evaluations and other forms of performance
measurement and assessment support the decision-making of government.
For government, and this department, the continual questioning of how we are performing is a critical
part of good performance management and accountability. We need to know:

Have we achieved what we set out to do?

How are we progressing in achieving the department’s strategic objectives?

Could we have done things better?

Should we continue to do this or do something else?
Through asking these types of questions we gain an understanding of what works and what doesn’t
work and why, what is being done well and what is not, what should be pursued and what should not.
This knowledge can improve the design and implementation of effective interventions.
The Public Governance, Performance and Accountability Act 2013 (PGPA Act) establishes a core set
of obligations that apply to all Commonwealth entities. The Enhanced Commonwealth Performance
Framework brings an increase in external scrutiny, and introduces new requirements for strategic
planning, measuring and assessing performance, evaluation and reporting.
Reflecting the department’s response to the PGPA Act, the principles outlined in this Evaluation
Strategy (the Strategy) will strengthen evaluation and performance measurement capacity in the
department and support building a culture of evaluative thinking, ultimately leading to better resource
allocation and decision-making and the evolution of programmes.
This Strategy:

Outlines the department’s approach to performance measurement and reporting, according to
good evaluation practice

Establishes a protocol for policy and programme areas to plan for evaluation across the
lifecycle of a programme

Introduces a strategic, risk-based, whole-of-department approach to prioritising evaluation
effort, and illustrates how evaluations may be scaled based on the value, impact and risk
profile of a programme

Describes how evaluation findings can be used for better decision-making

Describes how the department is building evaluation capability and a culture of continuous
improvement

Outlines how the department will measure its progress in implementing this Strategy.
This Strategy is not intended to be a complete guide to evaluation and performance measurement. It
is supported by a range of internal and external resources including:
1

the department’s comprehensive guidance material and templates for planning and
conducting an evaluation

the department’s Performance Measurement and Reporting Framework

the Department of Finance Enhanced Commonwealth Performance Framework

the Australian National Audit Office Best Practice Guide—Successful Implementation of
Policy Initiatives.
Performance measurement and reporting
The department’s performance measurement and reporting framework supports the implementation
of the Enhanced Commonwealth Performance Framework (the Framework) under the PGPA Act. 1
The Framework enables Commonwealth entities to devise the necessary links between their
performance information and their external reporting. Entities are encouraged to adopt performance
measurement methodologies that better assess the results of activities and articulate their
performance story. The framework introduces a more transparent and cohesive form of performance
reporting related to the activities of an entity in achieving its purpose.
Under the PGPA Act Commonwealth entities will be required to produce and publish annually:
A Corporate Plan, which sets out the purpose of the entity and the method of
measuring and assessing the entity’s performance in achieving its purpose.
An Annual Performance Statement, which measures and assesses the entity’s
performance in achieving its purpose in the reporting period. This statement is
published as part of the Annual Report.
There is a strong emphasis on performance monitoring, evaluation and reporting. Further, the
Framework establishes a clear cycle of planning, measuring, evaluation and reporting of results to
Parliament, ministers and the public. The revised reporting framework allows flexibility in using
various data sources to better assess the results of government programmes. Performance measures
are tailored specifically for each programme so that they reflect its design, including its specified
objectives, inputs, outputs and outcomes.
Traditional Key Performance Indicators (KPIs) can be complemented with other measures which
provide a more meaningful link between public resources used and the results delivered. These may
include benchmarking, surveys, peer reviews, and comprehensive evaluations.
Given the department’s high level of complexity, a hierarchy has been adopted to measure and report
on performance. Performance measures are identified and reported on at the level of the ‘Activity’,
‘Strategic Objective’ and ‘Entity’, reflecting the outputs, impacts and outcomes of the department’s
activities. This approach means that performance monitoring and assessment is undertaken at the
appropriate levels, where useful performance measures can be developed and quality data is
available.
1
Information in this section is based on guidance available as at 5 June 2015. Department of Finance
(2015), Enhanced Commonwealth Performance Framework, Public Management Reform Agenda,
webpage viewed 5 June 2015, https://cfar.govspace.gov.au/legislation-pgpa-act/#transitional.
2
Figure 1: Reporting hierarchy
█
Vision (outcomes)
Enabling growth and productivity for globally
competitive industries
█
Strategic objectives/purposes (impact)
Supporting science and commercialisation
Growing business investment and improving capability
█
Activity (intermediate output)
Business research, development and commercialisation
Economic transition
█
Component (output)
Entrepreneurs’ Infrastructure Programme
Cooperative Research Centres
The department sets out its vision and four strategic objectives in the Strategic Plan 2015-19.
Impact on evaluation activity
Good performance information will draw on multiple sources that offer different perspectives on the
achievement of a programme’s objectives. The performance story of a programme is likely to be best
supported through a diverse set of measures.
Evaluations provide a balanced performance story through their incorporation of programme logic
tools, and assessment against set criteria. They provide meaningful information and evidence on a
component’s aim and purpose in terms of its effectiveness and efficiency and the activities that
focussed on that purpose. They provide an opportunity to look beyond performance monitoring and
reporting and consider how well the programme is achieving its outcomes.
The department responds to growing demand for evidence-based analyses of policy and programme
impacts by applying robust research and analytical methods, both quantitative and qualitative, to
determine and isolate what works in industry and science policies and programmes.
3
What is evaluation?
Evaluation is an essential part of policy development and programme management. The continual
questioning of what we are trying to achieve and how we are performing enables us to learn and
improve what we do, ensuring that decision-making is informed by the best available evidence.
Policy and programme evaluation involves collecting, analysing, interpreting and communicating
information about the performance of government policies and programmes, in order to inform
decision-making and support the evolution of programmes.
Evaluation helps to answer questions such as:

Is the policy contributing to the intended outcomes or any unintended outcomes?

Are there better ways of achieving these outcomes?

What has been the impact of the programme?

Is the policy still aligned with government priorities, particularly in light of changing
circumstances?

Should the current programme be expanded, contracted or discontinued?

Is there a case to establish new programmes?

Can resources be allocated more efficiently by modifying a programme or mix of
programmes?2
Evaluation is integral to continual improvement. It is a not a one-off, or ‘tick the box’ exercise.
Evaluation supports:
Evidence-based
Policy Development
Public
Accountability
Learning
Performance
Reporting
2

evidence-based policy development and decision-making

a stronger basis for informing government priorities

more efficient resource allocation.

the public accountability requirements of programme sponsors and
governments

the department’s risk-management processes, helping to encourage greater
public trust in government.

shared learning to improve policy development and programme design and
delivery

a culture of organisational learning within the department.

the analysis and assessment of balanced and meaningful performance
information to report on progress in achieving strategic outcomes

an enhanced ability to achieve government priorities.
Davis G & Bridgman P (2004), Australian Policy Handbook, Allen & Unwin, Sydney, pp.130-131.
4
Good evaluation practices – a consistent approach to evaluation
If evaluations are to be valuable to decision-makers across government, consistency in approach
and planning are required. Evaluations should be conducted to a standard that ensures the
information is credible and evidence-based.
The table below outlines the key principles used to guide evaluation in this department.3
Evaluation Principles:
1.
2.
3.
Integrated
Fit for purpose
Evidence-based

Evaluation is core business for the department and is not simply a
compliance activity

Evaluation planning is undertaken at the New Policy Proposal Stage
or early in the design of programmes

Evaluation results are communicated widely and inform decisionmaking and policy development.

The scale of effort and resources allocated to an evaluation are
proportional to the value, impact, strategic importance and risk profile
of a programme

The evaluation method is selected according to the programme
lifecycle, feasibility of the method, availability of data and value for
money.

The department applies robust research and analytical methods to
assess impact and outcomes

Collectors of administrative data strive to attain baseline
measurements and trend data in forms that are relatable to external
data sets.
4.
Specific and timely

Evaluation planning is guided by the timing of critical decisions to
ensure sufficient bodies of evidence are available when needed.
5.
Transparent

All evaluation reports are communicated internally unless there are
strong reasons to limit circulation

The department will move towards publishing more content
externally to strengthen public confidence and support public debate.

Evaluation governance bodies have some independence from the
responsible policy and programme areas

Typically, evaluators will have some independence from the
responsible programme and policy areas.
6.
3
5
Independent
Adapted from Department of the Environment (2015), Evaluation Policy, Canberra, p.7.
Governance
Executive Board
The Executive Board is responsible for oversight of the department’s evaluation activity, including:

Endorsing the department’s annual evaluation plan

Agreeing the evaluations to be listed and reported on in public documents, including the
Corporate Plan and Annual Performance Statement

Observing progress against the annual evaluation plan

Noting emerging themes from completed evaluations and audits to help inform future policy
and programme design

Observing information sharing between the Audit Committee and Evaluation Unit regarding
evaluation activity within the department, including the implementation of major
recommendations

Considering whether to publish evaluation reports or summaries to support greater
transparency of the department’s performance and dissemination of learning as noted in the
National Commission of Audit.
Steering Committees and Reference Groups
Evaluations conducted by the department are overseen by a Steering Committee or Reference
Group. A Steering Committee is more appropriate for larger or more complex evaluations, involving
multiple stakeholder groups or layers of government. Steering Committees increase transparency,
cross-fertilisation and independence. A Steering Committee gives direction to the team conducting the
evaluation and ensures the evaluation is of an acceptable quality and is independent. While it is
beneficial to have the knowledge of policy and programme area representatives, there should be a
level of independence in the membership of Steering Committees.
A Reference Group differs from a Steering Committee in that it is a sounding board for ideas, but
does not formally approve or direct the evaluation. The Chair of a Reference Group should be
independent from the policy and programme areas.
Evaluation and Audit—what is the difference?
The roles of evaluators and auditors are quite different but they have similar outputs—both inform
performance reporting and address public accountability requirements. There are strategic linkages
and synergies between the two functions. When planning an evaluation or audit, and identifying
lessons learned to inform policy development and programme delivery, evaluation and audit activity
should be considered and coordinated. The responsibilities of the Assurance and Audit Committee
and the role of Internal Audit and the Evaluation Unit are outlined below to help identify who does
what.
Assurance and Audit Committee
The Assurance and Audit Committee—established in accordance with the PGPA Act—provides
independent advice and assurance to the Executive on the appropriateness of the department’s
accountability and control framework, independently verifying and safeguarding the integrity of the
department’s financial and performance reporting.
The Annual Audit Plan provides an overview of the delivery of Internal Audit services, which include
General audits, ICT audits, Management Initiated Reviews and Assurance Advisory Services.
6
Internal Audit role
Internal audit provides an independent and objective assurance and advisory service to:

provide assurance to the Secretary that the department's financial and operational
controls designed to manage the organisation's risks and achieve the department's
objectives are operating in an efficient, effective and ethical manner

assist the Executive and senior managers in the effective discharge of their
responsibilities

improve the effectiveness of risk management, control and governance including
business performance

advise the Assurance and Audit Committee regarding the efficient, effective and
ethical operation of the department.
Evaluation Unit role
The Evaluation Unit are the authoritative source of advice on evaluation. The Unit’s role is to help
build capability in performance measurement and evaluation, and promote a culture of evaluative
thinking and continual improvement across the department. The Unit has responsibility for developing
and reporting progress against the department’s Evaluation Plan.
The Evaluation Unit are available to provide expert advice and guidance to divisions in planning and
conducting evaluations. They conduct evaluations on behalf of divisions as directed by the Executive
Board and can provide expertise as members of Steering Committees and Reference Groups.
The Evaluation Unit is responsible for…

Providing expert advice and guidance to line areas in planning and conducting
evaluations

Strengthening line areas’ capability to conduct evaluations

Developing new tools to support line areas and increase the rigour of evaluations
across the department

Conducting or participating in evaluations of selected high-priority evaluations

Providing advice as members of Steering Committees and Reference Groups

Promoting evaluation thinking across the department and sharing report findings

Developing the department’s four-year Evaluation Plan and reporting progress
against the Plan to the Executive Board.
The responsibilities of the Evaluation Unit are further described on pages 12 and 15.
7
Planning for evaluation across the policy and programme lifecycle
The decision to conduct an evaluation is strategic rather than simply routine. Decision-makers need
to give considerable thought to what they want an evaluation to address and when an evaluation
should occur.
Evaluation activity has different purposes at different points in the programme lifecycle. All policy and
programme areas need to consider evaluation requirements from the early policy and programme
design stage (ideally this occurs at the New Policy Proposal stage and in conjunction with the
Evaluation Unit). Planning for evaluation at this early stage helps to identify what questions need to be
answered and when, so that meaningful data can be collected to measure the programme’s outcomes
and impact.
Evaluation types
The department uses four key types of evaluation, grouping them under the categories of formative
(as the initiative is getting underway) and summative (once results can be observed). 4
Formative
Prospective evaluation—ensures that the programme is prepared for robust evaluations from the
outset using a programme logic model to refine the proposal and determine performance measures
and data sources.5
Post-commencement—also known as a post implementation review, these evaluations ‘check in’
soon after the programme has begun. This type of evaluation focuses on initial implementation,
allowing decision-makers to identify early issues regarding programme administration and delivery
and take corrective action.
Summative
Monitoring—these evaluations draw on performance information to monitor the programme’s
progress. They are usually suited to programmes which are at a “business as usual” patch in the
programme lifecycle, and the primary audience are people internal to the programme. The findings of
a monitoring evaluation provide an indication of performance, contributing to the measurement of the
department’s strategic objectives; a basis for future reviews; and an opportunity to test the
programme’s data sources to see whether they are providing the required performance information.
Impact—Impact evaluations are usually larger and more complex evaluations which allow for
assessment of a programme’s performance. Where possible they would test this against a
‘counterfactual’: they seek to compare programme outcomes with a prediction of what would have
happened in the absence of the programme, and may include research about programme alternatives
to allow comparison of results. They may involve a cost-effectiveness or cost-benefit analysis.
This type of evaluation tends to have more of an external focus—the primary audience are people
external to the programme. Impact evaluations usually aspire to find objectively verifiable results,
contributing to the measurement of the department’s strategic objectives and outcome.
4
5
Refer to glossary at Appendix A.
United States General Accounting Office (1990), Prospective Evaluation Methods, Washington D.C.
8
Figure 2: Evaluation stages in the programme and policy lifecycle
Evidence from
evaluations informs policy
and programme design
beyond the programme
being evaluated.
Policy development &
Prospective
evaluation
Programme Logic
Impact
Evaluations
Evaluation
Planning
Post-Commencement
Evaluation
Monitoring
Evaluations
Programme logic is used at the policy and programme development stage to clarify
the intent of a programme and assist in the development of performance measures,
key evaluation questions and data sources, including establishing baseline data.
One year into a programme a post-commencement evaluation can be conducted to
examine whether the programme is operating as intended.
At the 12 – 18 month mark the data collection and performance measures of a
programme can be tested through a monitoring evaluation, to see if the indicators
and data are presenting sufficient information to assess effectiveness and impact of
the programme. This would also give an assessment of performance against the short
term performance indicators.
Depending on the nature of the programme, after three to five years an impact
evaluation can be conducted using an appropriate methodology.
Evaluation readiness
To help policy and programme areas to plan for evaluation and performance monitoring, all
programme areas will complete a Performance Measurement and Reporting Template, or similar
documentation. The information will inform the policy or programme Evaluation Lifecycle. Completing
these documents and instilling a performance-measurement mindset will ensure the department’s
programmes are ‘evaluation ready’, i.e. key evaluation questions have been identified, and there is
enough evidence and data to enable an evaluation to be conducted.
9
Prioritising evaluation effort
A four-year evaluation plan
The department has a strategic, risk-based, whole-of-department approach to prioritising evaluation
effort to provide greater oversight in the allocation of resources to evaluation activity across the
department.
As required under the PGPA Act, the department’s Evaluation Plan covers a four-year period (over
the forward estimates). Elements of the Plan will be published externally by the department, including
in the Corporate Plan and Annual Performance Statement.
While noting the many benefits of evaluations, it is not feasible, cost-effective or appropriate to fully
evaluate every programme and policy.
The department prioritises evaluation effort and resourcing based on the criteria:

Total funding allocated for the programme

Internal priority (importance to the department’s and Australian Government’s goals)

External priority (importance to external stakeholders)

Overall risk rating of the programme

Track record (previous evaluation, strength of performance monitoring, lessons learnt).
The department’s Evaluation Plan (published internally) is developed in consultation with divisions
using the above criteria as a guide to how and when evaluations should be conducted. To reduce
duplication and leverage effort, the department takes account of audit and research activity when
developing its Evaluation Plan, adopting a whole-of-department approach to evaluation planning.
The Evaluation Plan uses tiers to identify evaluations of highest priority and strategic importance. See
table on page 12.

Tier one—evaluations of highest priority and strategic importance

Tier two—second order of strategic importance and may include methodologies to test the
performance monitoring of the programme.

Tiers three A & three B—programmes of lesser strategic importance, low risk, or a terminated
programme. Tier three A includes single payment grants.
Alterations to the evaluation plan
The Evaluation Plan is not cast in stone. The department recognises that circumstances change as
months and years pass. It is important that ongoing intelligence about programmes’ evolution, and
adjustments in Government policy-setting, inform the Evaluation Plan. Generally, the Plan is updated
annually but alterations can be made at any time.
10
Scaling evaluation
The scale of an evaluation should be proportionate to the size, significance and risk profile of the
programme (sometimes referred to as ‘fit for purpose’). This means that evaluative effort and
resources should not be expended beyond what is required to satisfy public accountability and the
needs of decision-makers.
The selection of an evaluation method should also take into account the programme lifecycle and
feasibility of the method, the availability of data and value for money. Evaluations should be
appropriate to the particulars of a given programme; they are not a ‘one size fits all’ arrangement.
Options could include light touch/desk reviews, grouping a suite of programmes into one evaluation,
sectoral reviews and research projects that focus on the effectiveness and impact elements of
programme performance.
The increased accountability and scrutiny under the PGPA Act further reinforces the critical role of the
Evaluation Unit as the authoritative source for guidance on evaluation. To ensure evaluative effort and
resources are used to maximum benefit, it is a requirement that divisions consult with the Evaluation
Unit when determining an appropriate scope and methodology for their evaluations.
Responsibility for conducting evaluations
Priority, scale and methodology will inform who will conduct an evaluation. Subject-matter or technical
expertise should also be considered, as should resource availability, time and cost. Options include:
an individual or small team; working collaboratively across divisions or other agencies; engaging the
Evaluation Unit to conduct the evaluation; working in partnership with the Evaluation Unit; or engaging
an external consultant or academic.
The policy owner is accountable for managing the evaluation with assistance from the programme
delivery team.
There should be a level of independence from the areas responsible for policy and programme
delivery. For evaluations of lesser strategic importance or terminated programmes this could be
through an independent member on a Steering Committee or Reference Group. Seconding
individuals from divisions separate from the policy and delivery areas is a viable option to provide
some independence, build capability and alleviate resourcing constraints.
11
Evaluation approach
The table below outlines the fundamental issues which should be considered in determining the scale
of an evaluation. There may also be times where Cabinet or central agencies determine the type of
evaluation and when it should be conducted. The Evaluation Unit can provide advice and guidance on
planning and conducting evaluations across all of the tiers.
Characteristics
of Programme
Likely
characteristics
of Evaluation
Tier One
Tier Two
Tiers Three A & Three B

Significant funding

Moderate funding


Highest risk

Medium risk
Relatively small
funding or single
payment grants

Strategically
significant


Low risk

May be Flagship
programme
New or untried
programme that
requires testing of
assumptions and or
data

Lesser strategic
importance

High public profile and
expectations

Not widely publicised

Similar to other
programmes that have
been subject to
evaluation activity

Informal process

Can be completed
internally

Limited data
requirements

Low resource
allocation

Limited consultation

Low profile release

Politically significant

Formal process

Extensive consultation



Evaluation Unit
role
High resource
allocation
Central agencies may
be involved
Wide public release

Medium level of
strategic importance

Moderate public
profile and
expectations

Greater level of data
collection and analysis

Multiple evaluation
points during the
development and
implementation

Regular process
reporting
Evaluation Unit should be
consulted on
development of
methodology/terms of
reference.
Evaluation Unit should be
consulted on
development of
methodology/terms of
reference.
Evaluation Unit should be
consulted on
development of
methodology/terms of
reference.
Evaluation Unit
involvement either
conducting evaluation or
represented on Steering
Committee.
Could be conducted by or
in partnership with the
Evaluation Unit or by
independent
person/team.
Reference Group should
have a level of
independence.
Independent evaluator
could be internal or
external to the
department.
Evaluator could be
Internal or external to the
department.
12
Lessons learned — taking advantage of completed evaluations
Policy making is a process of continuous learning, rather than a series of one-off, unrelated
decisions. Effective use of organisational knowledge in policy development enables policy makers to
learn from previous successes and failures to develop better policy. Policy and programme
evaluations provide the evidence base to inform best practice expenditure of public funding and the
development of policy. 6
Evaluations increase understanding about the impact of government policy, programmes, regulations
and processes, and are just one of the sources of performance information that help the department
to assess whether it is achieving its strategic priorities. Along with research and audit findings, the
outcomes from evaluations are a valuable resource; they support evidence-based policy and the
continual improvement and evolution of programmes.
Organisational learning uses past experiences to improve policy, recognising that the government
may repeatedly deal with similar problems. Developing a culture of organisational learning can make
an organisation more responsive to the changes in its environment and facilitate adaptation to these
changes. 7
It is expected that evaluation findings will be communicated widely across the department, particularly
to inform decision-making, with resulting recommendations acted upon routinely. It is also expected
that evaluation findings and emerging trends are captured, reported and communicated, and brought
to the attention of the Executive Board as appropriate.
To improve the sharing of evaluation findings and make them accessible across the department, all
evaluations commissioned or undertaken by the department will be accessible internally through the
Research Management System. The System provides significant insight to the approaches used to
design policy and implement departmental programmes.
6
7
Department of Industry and Science (2014), Policy Development Toolkit, Canberra.
Ibid.
13
Building capacity and capability
Building capacity and capability in performance measurement and evaluation is not limited to
technical skills and knowledge. Performance measurement and evaluation needs to be integrated
into the way we work and think.
Fostering a culture of evaluative thinking
As we are called to adapt to changing economic and policy environments, measuring how we are
performing and providing credible evidence becomes paramount. This cannot be achieved without a
shift to a culture of evaluative thinking and continuous improvement.
Organisational culture significantly influences the success of evaluation activity and requires strong
leadership. This department is building a supportive culture, led from the Executive, that encourages
self-reflection, values results and innovation, looks for better ways of doing things, shares knowledge
and learns from mistakes.
Without such a culture, evaluation is likely to be resisted, perceived as a threat rather than an
opportunity and treated as a compliance exercise.
To develop a culture of evaluative thinking the department requires:

a clear vision for evaluation and continuous improvement

clear responsibilities and expectations to empower staff, along with appropriate training and
guidance material

knowledge-sharing and tolerance for mistakes to encourage learning and improve
performance

a culture of reward to showcase effective evaluations

support for the outcomes of robust evaluation to build trust, welcoming the identification of
problems or weaknesses.8
Building evaluation capability
A culture of evaluative thinking and capability building go hand in hand—both are required to achieve
a high level of evaluation maturity within a high performing organisation.
Conducting an evaluation requires significant knowledge, skill and experience. The department is
committed to building performance measurement and evaluation capability and technical skills to
support staff in planning and conducting evaluations, and undertaking performance monitoring.
Learning continues for staff in the Evaluation Unit and across the department in specialised evaluation
techniques and methods. The Evaluation Unit is made up of evaluation professionals who are
members of the Australasian Evaluation Society (AES) and other professional organisations.
The role and responsibilities of the Evaluation Unit include building capability through providing expert
advice and guidance, and ensuring the department is meeting its external reporting accountabilities.
8
ACT Government (2010), Evaluation Policy and Guidelines, Canberra.
14
In addition to the points noted on page 7:
The Evaluation Unit can advise on the…

Development of programme and policy logics

Conduct of programme and policy evaluations
(prospective, post-commencement, monitoring and impact evaluations)

Preparation of:

evaluation lifecycles
(to guide evaluations throughout a programme’s lifecycle)

evaluation plans
(to guide a single evaluation)

Development of programme level performance measures (KPIs)

Application of departmental evaluation principles.
Supporting guidance material
The Evaluation Unit have developed comprehensive guidance material to support on-the-job learning.
The topics covered range from planning for an evaluation to how to conduct an evaluation or develop
Terms of Reference. The material is designed to be used in conjunction with advice available from the
Evaluation Unit.
The department also offers targeted learning on programme logic and developing performance
measures.
15
Measures of success
Evaluation maturity
Developing and maintaining evaluation maturity is an ongoing process that must be balanced with
other organisational objectives. This Strategy establishes a framework to guide the department
through the stages of maturity which encompass good evaluation practices.
To establish a baseline from which we can identify strengths and weaknesses and priorities for
improvement, the department has assessed its current evaluation maturity. While it is following best
practice in some elements of evaluation maturity, overall it is at the developing stage of maturity.
Maturity level 
Integrated
Beginning
Developing
(Current rating)
Embedded
(June 2016)
Leading
(June 2018)

Awareness of the
benefits of
evaluation is low.




Evaluation is seen
as a compliance
activity and threat.
Appreciation of the
benefits of
evaluation
improving.

Evaluation is being
viewed as core
business for the
department, not
simply a compliance
activity.
Evaluations
motivate
improvements in
programme design
and policy
implementation.

Demonstrated
commitment to
continuous learning
and improvement
throughout the
agency.

Department is
recognised for its
evaluation and
performance
monitoring
expertise, and
innovative systems
and procedures.



Fear of negative
findings and
recommendations
leads to a
perception of
‘mandatory
optimism’ regarding
programme
performance.
Insufficient
resources allocated
to evaluation
activities.


Evaluation and
performance
measurement skills
and understanding
limited, despite
pockets of
expertise.

A culture of
evaluative thinking
and continual
improvement is
introduced and
communicated
across the
department.
Skills in
performance
measurement and
evaluation
developed through
targeted training
and guidance
materials.
Evaluation website
and guidance
materials
developed.

The role of the
Evaluation Unit is
widely
communicated. Unit
seen as the
authoritative source
for advice.

Developing further
expertise in the
Evaluation Unit.
A culture of
evaluative thinking
and continual
improvement is
embedded across
the department, with
lessons learnt being
acted upon.

Evaluation is seen
as an integral
component of sound
performance
management.

General evaluation
skills widespread.

Improved skills and
knowledge in
developing quality
performance
measures.

Evaluation Unit
team members
have high order
skills and
experience which is
leveraged by the
agency.

Evaluation Unit
team members hold
and are encouraged
to undertake formal
qualifications in
evaluation and
related subjects.
16
Fit for
Purpose
Evidencebased
Specific and
timely


Frequency and
quality of evaluation
is lacking.

Guidelines for
prioritising and
scaling evaluation
activity are used.

Priority programmes
are evaluated.

Evaluation effort is
scaled accordingly.

Evaluations use fit
for purpose
methodologies.

Specialist and
technical skills well
developed to apply
appropriate
methodologies.
Data holdings and
collection methods
are insufficient or of
poor quality.

Planning at
programme outset
improves data
holdings and
collection methods.

A range of
administrative and
other data is used in
the assessment of
performance.


Developing skills
and knowledge in
applying robust
research and
analytical methods
to assess impact
and outcomes.

Robust research
and analytical
methods are used
to assess impact
and outcomes.
The department
continually develops
and applies robust
research and
analytical methods
to assess impact
and outcomes.

Evaluation and
performance
measurement
conforms to
recognised
standards of quality.

The department’s
approach to
evaluation and
performance
planning is seen as
the exemplar.

All programmes
assessed as
‘evaluation ready’.

Quality of
evaluations is
improving.

Effort and resources
are allocated in an
ad hoc and reactive
manner with little
foresight.

Evaluation activity is
coordinated. An
Evaluation plan is in
place and regularly
monitored.

Developing
performance
information at the
inception of a
programme is ad
hoc and of variable
quality.

Strategically
significant and risky
programmes are
prioritised.


17
Planning for
evaluation and
performance
monitoring is being
integrated at the
programme design
stage.

Evaluations conform
to agency
standards.

The department
employs strategic
risk-based, wholeof-department
criteria to prioritise
evaluation effort.
Evaluation Plans
are updated
annually and
progress is
monitored on a
regular basis.

Planning for
evaluation and
performance
measurement is
considered a
fundamental part of
policy and
programme design.

All programmes
have programme
logic, performance
and evaluation
plans in place.
All programmes are
assessed for being
‘evaluation ready’.
Transparent
Independent
9

Findings and
recommendations
held in programme
and policy areas.

No follow up on the
implementation of
recommendations.

Independent
conduct and
governance of
evaluations is
lacking.

Evaluations are
conducted and
overseen by the
policy or
programme areas
responsible for
delivery of the
programme.



Findings and
recommendations
viewed as an
opportunity to
identify lessons
learned.
Evaluation findings
and
recommendations
available in the
research
management
system to improve
the dissemination of
lessons learned and
inform policy
development.
There is an
improved level of
independence in the
conduct and
governance of
evaluations.

Findings widely
disseminated and
drive better
performance.

Website and
guidance materials
valuable resource
for staff.

Evaluation findings
and reports are
published where
appropriate.

All evaluations
include a level of
independence.

Findings are
consistently used to
optimise delivery
and have influence
outside the
department.
Reviewing the Evaluation Strategy
This Strategy will be reviewed in 12 months (June 2016) to assess whether it is meeting the needs of
the department. The measures of success will include that it is:





Consistent with the PGPA Act
Efficiently allocating evaluation effort
Leading to more effective conduct of evaluations
Fostering a culture of evaluative thinking
Ultimately contributing to more effective programmes.
Results of the review will be communicated to the Executive Board. The review will include an
assessment of the department’s level of evaluation maturity, one year on.
The evaluation maturity table is adapted from:
ACT Government (2010), Evaluation Policy and Guidelines, Canberra, p.17.
18
APPENDIX A – Glossary
Term
Definition
Activity
A distinct effort of an entity undertaken to achieve specific results or aim.
An entity’s purpose may be achieved through a single activity or multiple
activities.
Baseline
Information collected before or at the start of a programme that provides
a basis for planning and/or assessing subsequent programme progress
and/or impact.
Component
A programme or initiative within the department.
Cost Benefit
Analysis/Cost
Effectiveness Analysis
Evaluation of the relationship between programme costs and outcomes.
Can be used to compare different programmes with the same outcome
to determine the most efficient intervention.
Counterfactual
A hypothetical statement of what would have happened (or not) had the
programme not been implemented.
Evaluation
A systematic assessment of the operation and/or the outcomes of a
programme or policy, compared to a set of explicit or implicit standards,
as a means of contributing to the improvement of the policy or
programme.
Expenditure Review
Principles
The principles used to review Australian Government activities and in
the assessment of proposals for new government activities. They
include: appropriateness, effectiveness, efficiency, integration,
performance assessment and strategic policy alignment.
Formative evaluation
Generally undertaken while a programme is forming (prior to
implementation), this type of evaluation is typically used to identify
aspects of a programme that can be improved to achieve better results.
Impact evaluation
Impact evaluations are usually larger and more complex evaluations
which allow for assessment of a programme’s performance. Where
possible they would test this against a ‘counterfactual’.
Key Performance
Indicator
Quantitative or qualitative variables that provide a reliable way to
measure intended changes. Performance indicators are used to observe
progress and to compare actual results with expected results.
Monitoring evaluation
Monitoring evaluation draws on performance information to monitor the
programme’s progress. Usually suited to programmes which have hit a
“business as usual” patch in the programme lifecycle, the findings
provide: an indication of performance, contributing to the measurement
of the department’s strategic objectives; a basis for future reviews; and
the opportunity to test the programme’s data sources to see whether
they are providing the performance information required.
Outcome
A result or effect that is caused by or attributable to the programme.
Output
The products, goods and services which are produced by the
programme.
19
Programme logic
A management tool that presents the logic of a programme in a diagram
(with related descriptions). It links longer-term objectives to a
programme’s intermediate and shorter-term objectives. The programme
logic is used to ensure the overall programme considers all the inputs,
activities and processes needed to achieve the intended programme
outcomes.
Post-commencement
evaluation
Post-commencement evaluations focus on initial implementation,
allowing decision-makers to identify early issues regarding programme
administration and delivery and take corrective action.
Prospective evaluation
Prospective evaluations ensure the programme is prepared for robust
evaluations from the outset, using a programme logic model to refine the
proposal and determine performance measures and data sources.
Purpose
Summative evaluation
In the context of the PGPA Act, the strategic objectives, functions or
roles of an entity, against which entities undertake activities.
Summative evaluations generally report when the programme has been
running long enough to produce results, although they should be
initiated during the programme design phase. They assess positive and
negative results, as well as intended and unintended outcomes. This
form of evaluation is used to determine whether the programme caused
demonstrable effects on specifically defined target outcomes.
20
Download