The Role of Leadership in Performance Management

advertisement
THE ROLE OF LEADERSHIP IN
PERFORMANCE MANAGEMENT
Donald P. Moynihan,
La Follette School of Public Affairs,
University of Wisconsin-Madison
Presentation to Chicago Federal Leadership Forum
Have you encountered?




Strategic planning
Performance measures
Performance contracts
Pay for performance
The role of leadership
•
During my 20 years in the private sector as a CEO and
advisor to CEOs, I found that leadership, measurement,
and a motivated workforce create the foundation for
good performance. I am confident that the same is true in
government

Jeff Zients – Chief Performance Officer, 2009
Outline






Defining terms
Era of governance by performance management
From Bush to Obama
How do we use performance systems?
What fosters use of performance data?
Summary points
Defining terms
Performance management

A system that generates performance information
through strategic planning and performance
measurement routines, and connects this information
to decision venues,
Performance regimes

Performance tools create unprecedented pressure
on public actors to perform, in a context where
performance is defined by quantitative indicators
Purposes of Performance Information




Promote: How can I convince political actors,
stakeholders and the public that my agency is doing
a good job?
Celebrate: What accomplishments are worthy of the
important ritual of celebrating success?
Learn: What is what working or not working?
Improve: What exactly should who do differently to
improve performance?
Purposes of Performance Information




Evaluate: how well is my agency performing?
Control: how can I ensure that my subordinates are
doing the right thing?
Budget: on what program’s, people, or projects
should my agency spend the public’s money?
Motivate: how can I motivate employees and
collaborators to improve performance?
ERA of governance by performance
management
Era of Governance
by Performance Management




The rise of a doctrine
Not new, but more influential than before
Must justify actions in terms of outputs and outcomes
Basis for holding new structural forms accountable
Doctrinal logic for change
Government Performance
and Results Act 1993

Mandated:
5
year strategic plans, updated every 3 years
 Specific goals and objectives
 Annual performance reviews and plans
From Bush to Obama
Bush approach


Presidents Management Agenda
“everyone agrees that scarce federal resources
should be allocated to programs that deliver
results”
Wanted to integrate performance data into budget
process
Congressional Justifications


Center around performance goals
Pushback from Appropriations Committees
 Veteran’s
Administration told; “to refrain from
incorporating ‘performance-based’ budget documents”;
later told: “If the Department wishes to continue the
wasteful practice of submitting a budget structure that
will not serve the needs of the Congress, the Congress
has little choice but to reject that structure and continue
providing appropriations that serve its purposes.”

Two budgets required
Congressional Justifications


Department of Transportation told: “agencies are directed to
refrain from including substantial amounts of performance
data within the budget justifications themselves, and to instead
revert to the traditional funding information previously
provided. Performance-related information may be submitted
under separate cover.”
Negative consequences were promised for agencies that
ignored this directive: “If the Office of Management and
Budget or individual agencies do not heed the Committee’s
direction, the Committee will assume that individual budget
offices have excess resources that can be applied to other,
more critical missions.”
Program Assessment Rating Tool (PART)

5 year summary by OMB of evidence on program
performance for 1016 programs
 18
percent are Effective
 31 percent are Moderately Effective
 29 percent are Adequate
 3 percent are Ineffective
 19 percent are Results Not Demonstrated
PART as Evidence-based Dialogue






Third-party program review with a clear opinion
Greater emphasis on performance
The standard of proof for program performance can only
be satisfied by positive evidence of results
The burden of proof for performance rests on agencies
Entire programs are evaluated on a regular basis
The routine nature of PART creates an incentive to engage
Obama: A Pragmatic approach

“The question we ask today is not whether our
government is too big or too small, but whether it works
-- whether it helps families find jobs at a decent wage,
care they can afford, a retirement that is dignified.
Where the answer is yes, we intend to move forward.
Where the answer is no, programs will end. And those
of us who manage the public's dollars will be held to
account, to spend wisely, reform bad habits, and do our
business in the light of day, because only then can we
restore the vital trust between a people and their
government”
Example: Pedometer challenge!


Voluntary
Belief that transparent performance numbers will
change behavior, create a sense of competition and
raise performance
Early evidence on Obama

Performance measurement will be important
 “The
President is creating a focused team within the
White House that will work with agency leaders and the
OMB to improve the results and outcomes for Federal
Government programs while eliminating waste and
inefficiency”
 Chief performance officer
 Continue to maintain agency level performance positions
What happens to PART?




Not clear
Criticized as ideological, as too broad, as a data
collection exercise
Analysis remains in place, but new PARTs have not
started
OMB have offered agencies funds for better
evaluations
New emphasis on leadership



Focusing leaders on what matters – key goals
Accelerating results – Performance Improvement
Council; data driven meetings
Style: focused collaboration
New focus on information use



Will be a central aspect of the Obama
administration’s performance initiatives
Jeff Zients: “The ultimate test of our performance
management efforts is whether or not the
information is used”
Shelly Metzenbaum: “the key performance
management challenge facing the Obama
administration is to use—not just produce—
performance goals and measures”
How do we use performance
systems?
Why care about use?





For reforms to succeed, implies that data is used
Provides a tractable means of studying the impact of
results-based reform
Public organizations have devoted significant time
and resources into creating routines to collect and
disseminate data
Almost no attention to creating routines of use
How do you use performance data?
Types of responses: 4 Ps




Passive
Perverse
Political
Purposeful
Passive use of data

Passive:
 Do
the minimum to comply with requirements
 Do not actually use data
 Correlated with cynicism about reforms
Perverse use of data




Effort Substitution: Reducing effort on non-measured
dimensions
Cherry picking/Cream-skimming: Focusing effort on
subgroups of clients most likely to provide greatest
impact on performance measures while effectively
denying services to others.
Measure selection: Selecting metrics or data to measure
that will offer the most favorable portrayal of a service
Hiding numbers: Declining to present performance
measures that may exist
Perverse use of data




Output distortion: Manipulating measurement
processes to improve measured performance.
Ratchet effects: Curbing productivity in one time
period to avoid the setting of more challenging
targets in another.
Churning: Frequently adopting different targets or
measures to prevent comparison across time.
Cheating: Simply making up numbers, though rare,
does occur.
Responding to perversity

Add new/additional measures

Change existing measures

Rely/cultivate intrinsic norms to limit misbehavior

Avoid high-powered incentives
Political uses of data


Process of selecting measures means shaping a
program narrative
“Understand that measuring policy is not a science.
It is an art. It is words, and pictures and numbers.
And you create impressions, beliefs, understandings
and persuasions.”
Political uses of data


Data tells us what happened
Program officials still need to interpret and explain:
why performance did or did not occur;
 the context of performance;
 how implementation occurred;
 an understanding of outside influences on performance; and
 how to choose which program measure is a priority.


Exploit ambiguity and subjectivity of data
Political: Ambiguity of data



Examine same programs, but disagree on data
Agree on data, but disagree on meaning
Agree on meaning, but not on next action
steps/resources
Political: Subjectivity of data

Actors will select and interpret performance
information consistent with institutional values and
purposes
Evidence of Ambiguity in PART

Ambiguity of terms:


E.g.: Program purpose, quality evaluation, ambitious, having made
progress
How to interpret results? Multiple logics from experiment:







Argue that ratings are unreliable
Cut poorly managed programs
Raise funding for programs with positive assessments
Parity: Raise funding because program with similiar assessment received
more
Delay cuts because progress being made
Clear relationship between resources, need and program delivery
Stakeholder and congressional views
Evidence of Subjectivity with PART

OMB using PART to expand influence in
performance management/policy
 OMB
can define programs, goals, measures, agency
responsibility


Disagreement with agencies/Congress on
meaning/relevance of PART
Experimental evidence:
 UW
students significantly more likely to disagree with
OMB, and to argue for higher assessments and
resources
Implications for Decisionmaking



Performance information use reflects political
process, does not replace it
Performance information use does not lead to
clarity
Ability to structure dialogue tied to power
Purposeful use of data


Use data to improve program performance
Goal-based learning
 efficiency
improvements
 better targeting of resources
 more informed strategic decisions,
 tying indicators to rewards/sanctions in contract
arrangements
Purposeful use of data

Use of performance information for problem-solving
more likely to occur in intra-institutional settings
 Reduces

competing interpretations
Problem of neglect
 rarely
do anything with information
Learning forums



Routines specifically focused on solution-seeking,
where actors collectively examine information,
consider its significance and decide how it will
affect future action
What measures are useful for agency officials?
What other ways can we encourage learning
forums?
What fosters performance
information use?
The Right Context




Simple function that is easy to measure
Clear link between measures of actions, and
measures of outcomes
One-dimensional – relatively few measures that do
not conflict with one another
Stakeholder support – clear agreement about
purpose
Other factors





Learning forums
Mission-based culture/supportive culture
Resources
Administrative stability
Administrative capacity
Quantitative approach



3 studies using survey-based data
Self-reported performance information use
Results from Moynihan and Pandey (in press) and
Moynihan, Wright and Pandey (2009; 2010)
Study 1: Ordinal regression of reported performance information use for decisions
Variable
Hypothesized
direction
Result
Individual beliefs
Public service motivation
Positive
***
Job attributes
Reward expectation
Generalist leader
Task-specific experience
Positive
Negative
Positive
-***
--
Organizational factors
Information availability
Developmental culture
Flexibility
Budget staff take adversarial role
Positive
Positive
Positive
Positive/negative
***
***
*
--
External factors
Citizen participation
Professional influence
Positive/negative
Positive
+
+
*** = significant at .001; ** = .01; * = .05 +=.10 (two tailed tests)
Controls: region, income per capita, government size, population size, population homogeniety
Intrinsic vs. extrinsic motivation
Sense of public service motivation mattered
 Possibility of extrinsic reward did not create
an incentive to use data
 Implication: performance information use as
extra role behavior

Organizational factors

Information availability
 Supply-side
approach
 Use increases with better information, and when
information is tied to management systems
Organizational factors

Demand side approach
 Culture
matters
 Previous
work focuses on whether culture welcomed
performance management reforms
 What about broader measures of culture?
 Developmental culture (adaptability, readiness, growth)
 Flexibility
– unlikely to use data if cannot apply insights
Specialist vs. generalist leaders


Task-specific knowledge provides context in which
to interpret and apply data
Leadership role
 Task-specific
leaders more likely to use data than
generalist leaders
Other evidence of leadership
Support/commitment
 Provision of resources
 Participation


What other ways can leadership matter?
Study 2: Transformational leadership

Approach to leadership consistent with
performance:
 Articulate an appealing vision of the organization’s
mission and future
 Model behavior consistent with vision, inspiring role
model
 Challenge old assumptions
Propositions


Transformational leadership behaviors will have an
indirect, positive effect on performance information
use through its influence on goal clarity
Transformational leadership behaviors will have an
indirect, positive effect on performance information
use through its influence on organizational culture.
Key measures

Transformational leadership
•
Asked department heads/assistant city managers on extent
to which city manager demonstrates transformational
leadership:
 articulates his/her vision of the future.
 leads by setting a good example
 challenges me to think about old problems in new ways
 says things that make employees proud to be part of the
organization.
 as a clear sense of where our organization should be in five
years.
Aggregated responses by organization

Structural Equation Model
Goal
Clarity
0.66*
Transformational
Leadership
0.60*
0.57
0.26*
E
0.32*
0.33*
Performance
Information
Availability
0.52
0.13*
E
Performance
Information
Use
0.63
E
0.17*
Developmental
Culture
0.89
E
All paths reported as standardized coefficient
*p < 0.05
Implications



Leadership and management
Indirect effects are important
“Setting the table” as long-term leadership strategy
Study 3: Perceived social impact



Individuals who see their work as helping others
more likely to use performance information
Some evidence that individuals who perceive
greater social impact are more motivated
Why should it relate to performance information
use?
Key measures

Perceived social impact
I
feel that my work makes a positive difference in other
people’s lives.
 I am very aware of the ways in which my work is
benefiting others.
 I am very conscious of the positive impact my work has
on others.
 I have a positive impact on others in my work on a
regular basis
Purposeful and political use
Purposeful
 I regularly use performance information to make decisions.
 I use performance information to think of new approaches for doing old
things.
 I use performance information to set priorities.
 I use performance information to identify problems that need attention.
Political
 I use performance information to communicate program successes to
stakeholders.
 I use performance information to advocate for resources to support
program needs.
 I use performance information to explain the value of the program to the
public.
Table 2: OLS Regression of Performance Information Use
Independent variables
Purposeful Use
Political
Perceived social impact
.43 (.12)***
.52 (.11)***
Goal clarity
.19 (.10)*
.25 (.09)**
Centralization
.05 (.07)
.10 (.08)
Political support
.14 (.07)*
.06 (.07)
Red tape
.04 (.03)
.02 (.02)
Years in position
.01 (.007)
.006 (.008)
Supervisory level
.03 (.05)
.02 (.04)
Educational level
.05 (.07)
.09 (.06)
Sector
-.06 (.15)
.08 (.14)
*** = .001;**= .01; *=.05 one-tailed test
N=184; R2= .25
N= 186; R2= .31
Study 4: Experimental approach





How does performance information matter to
decisions?
How does the framing of performance information
affect decisions?
Respondents given surveys with scenario – make
budget recommendations
Series of vignettes for different programs
Half vignettes are control, half are treatment
Theoretical background


Research on decision frames from psychology and
behavioral economics
Performance information is strategically selected
and presented – does this work?
Does the Addition of Performance Data Matter?


Control: no data; treatment: addition of data without
clear correlation to resources
The Department of Land and Water Resources is responsible for monitoring
and maintaining the water quality of lakes in the county, including two major
lakes that are popular for swimming and other water sports during the
summer. Estimates of water quality are based on pH levels, pesticides,
nitrates and other chemicals in the water.
2007
2008
2009
Program funding
174,000
179,000
182,000
Number of days water quality of major lakes
deemed unsafe
12
14
11
Does the Addition of Performance Data Matter?


Control: no data; treatment: addition of data with
clear relationship to resources
The Department of Social Services delivers a program called the Home
Downpayment Initiative. Using a mix of federal, state, and local resources,
the program seeks to increase the homeownership rates among low-income
and minority initiatives. To do so, it provides financial assistance to firsttime homebuyers for downpayment and closing costs.
2007
2008
2009
Home Downpayment funding
723,000
747,000
769,000
Number of families purchasing homes
36
38
46
Is outcome data more powerful than output?


Control: output data; treatment: outcome data
The Department of Health Services offers a program called Health Check,
which is a preventive health check-up program made available for anyone
under the age of 21 who is currently enrolled in Medicaid. Health Check
provides a head-to-toe medical exam, immunizations, eye exam, lab tests,
growth and development check, hearing check, nutrition check, and teen
pregnancy services. The goal of the program is to prevent the incidence of
more serious and more expensive health situations.
2007
2008
2009
Health Check funding
232,000
244,000
269,000
Clients treated
1232
1401
1325
Estimated savings due to preventive care
383,000
402,000
389,000
Threshold effects

Treatment: performance data pass a memorable threshold
(200)
The County Tourism Board seeks to increase visits from those who live outside
the county, and to increase the use of recreational and cultural opportunities
by both locals and outsiders. It collects data from local hotels, restaurants,
and other businesses that depend on tourists. In the last number of years,
the number of tourists visiting the county has stayed relatively flat at about
100,000, and the Board has focused its marketing budget on “quality, not
quantity,” by increasing the dollar amount that each tourist spends.
2007
Program budget
Average daily dollar amount spent by tourists
Program budget
Average daily dollar amount spent by tourists
120,898
178
2008
124,490
184
2009
131,289
195
2007
2008
2009
124,294
127,196
135,329
183
188
201
Including equity measures


Treatment: addition of equity measure that aligns with mission
The Department of Social Services funds the Early Intervention Program, which
provides services for children three and under with developmental delays and
disabilities. The mission statement for the Early Intervention Program is: “Our
mission is to provide access to therapies that improve child developmental
outcomes.” The program is administered by a non-profit, and employs
therapists to work with children and families in the home environment.
2007
Program budget
Number of hours of contact per year
Number of children served
Percent of eligible children below the
poverty line that access services
2008
2009
329,677
333,451
341,386
8901
8977
9011
212
233
221
42%
47%
51%
Summary points: what to do


Move beyond passive and limit perverse use
Focus on political use
 What
is the narrative of your program?
 What goals are meaningful and telling? How do they
relate to the narrative?
 What goals are essential to explaining program
purpose and achievement?
 How do you frame and communicate measures? Who is
your audience?
Summary points: what to do

Focus on purposeful use
 Provide
resources, be involved, make clear that it is
important
 Encourage right context for use
 Foster
goal clarity
 Encourage supportive culture
 Create
and support learning forums
 Appeal to intrinsic motivation
 Focus
on demonstrating significance of measures
Questions/Comments



dmoynihan@lafollette.wisc.edu
http://www.lafollette.wisc.edu/facultystaff/moynihandonald.html
The Dynamics of Performance Management

Georgetown University Press
Download