Document 11118055

advertisement
-II-
.;--I-I-..:II
--I711.I....
..,:,....I;p-.:,-II-...
Z.I-.1.,.-... I--:-,-;--..1,-.i1."-.-'-..-I.I;-i..711.,i..
., .1I..
-.
I -.I-I.II....--.-..-i,-I--I-I..-I1.--1,,....
I...-,I-..-I:-o-I.....-,-I-o.-.--,,,--...-1.-I.
-- ,,,I
-I
-.I,
:.,.I..--.,-,--.-I.--.
:,.--1:I--_:,.-:".--,.,..I-iI:I
,I-...
,I
I,.
- o, f,-I,
mIII-1--:,
,,:Z.I.1,I.-,.
I,--.
..
I.-.--..,
-I.--11.-.-,I,.-,,-,-.I:I ,III:-.,, ..II-I.-.,-I
,; "., -I-1,-----.,o..I..,
.--...,--,-,. II---.,.,
-,,:.:
-,Z,.
,,-i,.,
.- ...::,,i,,I.-..-.--.-.-I.,-.
,--.i"-.:..,I:.I
-,.,II,.I
,-,-- :,.:,-.I:--,.-,-..--,-.-.-,,,-:,-,,2,
I,.-.-..--,-,,-- :11.II.-.:
I,
::I,...-.I.
."-.
.,.--I,...,j,-...-::-.--,---.,--,,,
---I,.,----II---1..
---..
"'.
:1
.",
,
,
---:
-,,,
,--,.,,,,,,--Z.-.,,,-..-,-,:.;;,
;,.%.,,
..
,-.,,
-.
.-.
,1,
,.1-,; ,-.- i -,
,-I
..,--,,--.
,!.---I.1,.-.:,.,..--.I;-,,
,,,."-,-;.1 -,-,..-:,
.- --, ,,-,. I-I- . .,- .-,-I..1 ..-.
,!-;
--.
. ,.,--.-I,.-,
II------..rl.:,--,,
-,:
F,
a", .,.--1
""';, "
:--- , ,7-,-I
II.--.- -1
,,I .
:-'.----,
"
_,t--.
-,
,---,
,-,-I.%
,,--,..,.,.:-I-,. ;,,
IIt,1.I.
i,
I:, .
..-I. .----4,.-1,,-I.,,:,:-..,
-.
II--.--.,,,,..--i-:.e.-.,,-. ,.,,I.-.
:..I -.I--j.-,.-.,-,-,,,
,2,.-,
,, ,-', .- -,- - I-: ,
-1- -:,
,I. -. ,.,-,.I,,,,I: ,,. ,m , '.
-%
, . . -.
: I
:
-.,
;,.
I
I---- -I -,
I
, ".---.,:-,-,...,..I..-- II
.,..I. .
- --.
,.-,I,-, 4.-II.-.- .:--..- i
I
,,,-:z;,.,
,,,,.
-1
,,,,,: ,-,. ...I,-,.
....
-.
1-L,
-,.-. ",.lI,, -- --.---.
,--.-- ..
,
I....-III
-.4 :z-:,,
-,
, ,:.-,,.,,,.-,,..-,I--.-,.
-.w:...1,-,:,:.-.II,,-,.i---I- ..,,,II ,,-.,
.I---::.-.,.,,:.:;,--"-,,-,..,,
,,.I....-'ill
..
:.,,..
""
.,.I...
,,:-Z,:,I1,I.I-..-..: --........-,-,--,,,.
II- -t,7.
,z,
;. : --.
.
,;III..-.. .,I
I-.-.-', ,li -II,,-.
I,--,.,-,--,
----.
-,n-,
-:...11
-I
4.-,I:,.,
,-..I-I,.,I-I,-1,
,---,I;..q.
,...-..I
I.I,.,,.,..
.r ,-,-,
-..- .-"--,,..4,-.:-:-I--:--. -I.-.--II
..
,-,,.,.I..,.
: .: - -,--":1
,.--;
,.,,-- . ;---,,--, , ..,.-,.;:,,:
.,I,
,; ,,,,.i.,..-,,--,,"
,--,r-:.:
.- -7
,I1:
-,:`:,..
.,. .,,----,.-;,.,
,--"
..
:.:-.-,!-I,.
:---,
,:,
.,
,,1
,-,;-.,
,I-.
.;:.
-,-i"1:--:
.,
,., -.,,
.:,I.1:
,--,-,.,,--.
_-,-C-_-.
.1I.
!,
57
,-..-'
-i ,:;,,.-,,.-, -:i,`.---.--.1.,
;.
,--.-.-..,:,Z-;,
------`
-,
-,
-,
,-, :.;,-i--,,,..
-,
,:,:
--.-,,--,I.
...
::,,.,.,,,--,,I:.....,.--I,:....
, ----.
--.
7- -,,m-..-,I-.,,.
Ir..,--1
,,,-:,
--,I,-.-.-,.-,,II,--.I
I,--..II11
1,,.
,.::,----.,.I,
f,,I..,"-:-1 ----.-,,
--.. ----11,---4.,i,--7
,,,..--,
-,-.
I---,..,.,-,-,,-,,,-,
.",-I;-::,
,,-,!-.". ',,-,--II-,I
-I-..1..,--,
., ,:I.
....
:.m,-.-I,-q.IZ--.
I- .
,-`I.:-I--..
,-,
-;I.I..III1..--.
--.
.11
.-,!..--..;.--:.
,-.:-,-r:: ..:
.%.-.I-,,,,,--I,I%.I.-..,-;.I..-,..I
-',.'
::,...,-I, ,-II.--.
-.,
..: -..I,.,
: --, ,.
I . . ..
,,-..-,
I .:7
-.
I..!:I.--,I,I-II
,-I..,.--I-4.I---: .I-I.,- -- :.:.;I
,,-,,-,
1- :: ,-..-.-,,I-!.-.
----.--.IJ,-,- -- 1:
-I,.
II----,
,,.,-I:,---,.
II-. ..--- I
-.- '--- I:.:I..-I.II!,
-;-. .---,--.
,-,.II:,,,
III
,..,-q--,,,I-I.-.
,".--,.
...-. ---.--. .----.
--,
-- ,,.-,,
-:- !,
.
.";:
-, ,,, -,----'i,,
I. -,II:.
-..-.:,I1:,,--,-:i:,,,.-,I.--,:
-,.,m.:
,.,,-,:,.: ,-,:1
-`,
:.- ,
,
:
..
,----,.--I
.------,"'
.
:.
:--.:::.:-,.
,,o,--.II--,
-I-.
I.---,I--.-I,..
:I.
;
--:,-:IIP-.,-Il-:,
-'.,--I,:..I1.-I-,.-.II--i,
-1
I
.
J-.,1,I
,7 -1---.
. -,.I
I-..---.
I,
:
----,.
,
..
.-i ..I :.
.,o..-;,---"-,
1. -,.
,,
4.,I:: -`,--,-'.-,-,.
,--I
,-:t
,--,.,
,
,'.1:I,'.-,
-I-II
;
,I--,II--I.II
.
I-.:-,,
-.:-I
,,.-:,::Z
,-,I
-i
-,
.--:,-L,-:-,..
.,
,: -;
I:I..I
-I i.,.I,
,,
-.
, 7 , .. ,
.-- I,---!,,, ,-,I.:.:.-:
.- -.
.,.i
. I.I-----:,
:-.1;--.--.Z:
II-----,
IiII- .:cIII-,i,-,--I.I:"
:,:.----.
:- J,.:. -, A,,
P,--.-.,-,-,
-I-...I,
.,.,,
,
.-,
.
;I--.:I,II.I.-..:I-...1I-.
,-;1;-I.-.,,
I-.II
.,,I-;
1.
... ...
,
-I--,--rp-.II..--'-I-.,:--,-,,I.',
,-,I--,
: .:---,-.,:,% :-..
,7,:-,..,:I,.
. I:,I,,,-,dI,,
,.'", -,...--,.
-:1II---ZI..
,..-,
:..,
..
-,
-,,
,
.,
:
.,-:I.Z-II..-,--.--,.
.%
-.
:,
- -,.
-:. Z
-, .-.-i11,
-:,
.,,-,.-I-; .-7
-::-,
-I!-II
,I:,.',
I-.:..:;,
.,,j
.,::-,
I
.11
:1.;p.-,.
,.--:.
,--.-.i
.-..,
-,
,.
-;I
,:,-,
-,,.f ,-1%,
:..---,-,.,:--.
.
.,, ,.;,-..,.-I
,--.,-.V-.....I-,..
,,Z
,,-:.I
-,-..,I
I-I-,I,,,,III
.,
,--.
..,,,.,..11--.-I.
.,.II. ... --,--,-,
----,
-...
,,
-::-I.,,I-.,--,.-,
..
. .Z.
,, q,.,,.,.,
.,I,,
:. -:1,,7:..,.-.:,,,"
,,"..-,
I-..I.
-,i,:. :,.-i.-,,-.,,--".":-,,
:
,-:.,,-I,
,.v.
",
,I-"..-,:,
,,,,
.,- -:--, ,, : ,",
..,,-,,.
,-,-_:,,,.. 2,,,,,,..,-:,
,,--., ,
.-III!
.i-,-,,.;
._r,
I.,.
III :;,.-:..
,I:....-,-I
,;.--.-.,,-, ,.I -. ,.-;-,4-.,--.-.:
.1.-I-,
I.I ,
,,
,,,,I .: .,
-,,-- ,:."-.- ,. I
-. .---_-,
-,-,.-;,-.-.
--,-.I ,.-., ,..,,.,.,,Z.
, ",.",,,'.,.-,
,-J , ,,.-.,
:-"..,,.I,,--,-.
,,.II
:..--,
-,--1--.------,.---,
, ,- Z
F,-,--.- .t.-;',,--.-.,.,,.
---.:..-,,.,.-.. ,,:Z-I
":--..
I,-...: I .,,--I.II
I : I-.--I
1.-:
,,:,,IV--,.:
- ,.
,:-,;I-----I
I:..:,...
I,-.-,-....m:,.,.,-.., -,.-I"-.-I .- .,:,1.,--,.,I-I.....,
,.-- ..,`: :',::, :-I
,. .,,.
,1,- .,.,-.,.;,. -I--:
,.. .. , .7
-1. -;
-I..
: '.,:, 1,:,4, , I-..I..I,-..
a,
,-, :-..
-,
.II
I
,
;
11,II--I.,
.I-I
I-.-I",--;-.
,IT,",
,-.I-..1
-;'-;7-,.
I,.1 ..: -.-.
,..
II.
..11 I11I,,
----17:',
, smalk,--,: -,,-- l.
.i,..
,,t-..
...
,,,5i.-. ,
..-..-.:ID
-,,,-,
--1
: ..--..1I..r-W.
Is,:,,,. .
. -,-,:.I.-,, :%..w'r
,.I.
".
,---,
.---,---.,..,.:
:
.,,:
,I.,
-.
r%:. ,
1: -; : iI.-7,,-,..
".:,.:.,--.:-I:,.
: ..
r,,.I.r...
,r.,---:,-.--,.t
--.... :- L7--.
.,.--,
-.L
-_
.I.,- :-.--.
i,..,Ir,,,,. -,,-,
,.I.-.O
,.,` " -. - ,,- - ,-,-,,.--.,--,I.I
-- -. I..-,
1:.,.
",,,.,
--.,--!I
,,-:
r-, -,,-,-.
11..--,.11r,,I-.j,.--,,,:I,
-,:
,,",.I:.---,
,--,,,.-,-,.-.
: , ,:, ,,.I-,..,.-,-,--I-..
-.I-1.-..
II ,.;4;I
.,,"-11
.4--....,4Ii
1: ..
' ,.1.LII.,-I.I,;
,-I1ri
,,.
..
,,-v
;1,:::
,,-,
r : .__,:,I
.,,
,1,-,-I,.
--I -, ,.
.I._-..-,
, ". ,, i.
.1
,.I.IS.
: I,o.pI 4,--.i:I:Ib,.
,,-.. ,-: ......
4-,,7-..
,..
,.
.,-,
-.I:;,-,,
,--I , 1.,.-I,
..I---.I,
;--"
,.,-:::i
.,
:
,,,
,
:.1
.,.,..Ii:--,,-...
:.,
-I:14
.,."....:.-.
.,-'L-,
III--..,
,,,,--m11. ;:--I,1,.-Ir.I-.-.,, ,: ,,II
II.,I. I
I.,,-.-,,,..,
,,--,----, -,,.-,,:.,I.-,.:..
,-:-,.:,I...vII.-,.,
.--:i.--1
,,-1,I....IIr,,,,.I,: .I.:: ` - ,,:
:.,-.:
----I. ..
1.,--,..
,-,,-,--r:.--.I,7,.::,.---1.
.-...,:-.-.,:,
-..L...,
,.
::,,-III
,...1.I-,.--,,,
-,
I---I.I,-I-7.i-I.-..
,,,-I.
-,..
:--.-:. :",
,,,--I,.:-:,-:.
,
-...-.-,.:.
-,--::---,.-I-rL.
.--I:2.-I
I..--.-,..
.. ,-I , - I-----,,.-r..II...-.:- ..
,,.
--.
:-.
I--.
,:-.I-I-----,--:"
:. .,I-i...
,I,-I,,I..i;-:1.
-_---:r
.II;,I.
,,,.II,.,.I:--..,
Z
I
:1, - ':r,-'.'".I.-,.;,--.I.-.--I:..
m.,
: .."
II-..
:.,1.I.
-.
-1-..
,,,-11::
.,,.
.I.-I1.L.I
,,-,-.
,,,---,.--.--,I-._,_.L.I.%.--,
I.---,-411-''--,..
--.
,I-i--I-,
:,..",.1----:-,,I
q----I,I.,..I--:-.
...I,-I,.-..
:4----I
....
I%
--Z,-,,,..,,,.
:Lj,:I-L-.,II..I
.-.---,',.-7-,_-..-I-.I1
.I-P
,-:-.1
:-.,-4:'
1,
.- ;,-I..I.I_L
--L :1..I-I.,-.I..--I-.:-II-,--II,..Z
.-..
--.
L-I-.
. II:.'I--I- -L ,,--. ,..`I I-.
.,,..
.. Ir.I.,.:
.I. ,-1---._,
- .,-,_I
j-,-,IL:t-I--:-I1,,I-,--,-,-,.-L."L-.,.I
,,.I-II..-.--I'
,,'_I-.-,.-,,,%
'-. -,., : '
' :
: - ,.. -' . ,,
,I..-, . I , -.
-I
-,
,I
-,-,.'
,:
.
L
_
L.
L
,I--.,--,1-7
.
11-,
.L--.-,,.L
.
.I .
IL , .
- . ,-: -.
: ',-', - I I
-,-.- -,,
I
: r,: - . .--,
,r.,.:-,,-..Z:--.--,_
II.1.
-,..,,,-, :::-.-. ,,.,.,
:,
I-.,- -:,,
--1I.II---:-_.:-:L
,--_LI.
:.:: ,: -1 I-.
mII:I
i :,,--Z
:II.I-.74.,--..:i.
.-7.I1.-_,,.-.:,
L.,.----I-,___,.
.L.L-7.--7
:,I-,,.---I,-,I-,_..-,
.L.I--L,,-,_.,;:,
,I._,.
-.
._;I-.,,-,I.
,,.,
Ip-,
t;I,_,.,-:-L,..I,,r..-,-.
I,;-1..%-LI
,I;..I.---.
:1,.,I I .1-: .,. :I.I..:-;:`I--..LII,;I
LI--.%
-'-II11.-,.,.:
):...I-T.
,.-r,-.-,",_I-----,
:, --,,,'LL-,1.,-.
-I Ii
-.1L,.-L,..II
,: I.I-:.,Z..-,.
.! ..I:I-I--:-.I-i...-I-.I:,,L-I!---,-I"I
,: ,.
,L
,..::-.
..I- . .Ll
. .LL]::
:.-.,.-rZ.,
-,.:,..
,.-.I
-,
r,II,_.",,..II.1-':----:,
-:-,-..---:'...
L,-.-_..,-'l
L.-:-.,I.,:,.kl
-,-.
:,---.1.:
, ,
:,;.,,L','
-.
,:: ,-,..I---,:,
,-, ., -I I-: -..-,,, ;:;.--Ii
,I,.:II
:L...
:.iL1....--.I..-,,,.
7
,X.-'L,
:___,-I.
-1
i,
.I-.-..LII
-i -I.iL--L;,
.;.:I;
.-,
.: ,::,,Z,
'-I'-.,,I.,..-I-...L-,';,L .,_,
.,:--!"'-',.
.L.1..r.-L1
.-.
L,-,_.
_,,I
I._....
-: .
I-I,.'-.-.-,.
,.,-,:,I-L-.. -I.
l I.II-I,IL_.-j,:,.;L.-.,
I.I:.'.IwI..-_-1
II,-.-.
r.
Ii.-...
-i,:%
II:',
: ,
.:.-'...
:
II.
I..,,..L.-L.
-, _ .. ..
---,I
_-_, :'`,:I..
;,i: ,..,,--:-.I-:
,i,,. ..L,-c-:!:
I ,-L..,
.,.,I
II..IL.
I . :__ :1
:-,,-,I.,I-I.:--I-..-,Ii
-,I
,I; -'L'-:.--,,,,, IL--7.N-1
i";!.,I,-,-r..-...,I..;
I..-,I.-.,,,:.I
-I.
..
IITI Li,--,--I
,.,,
-I;I..,11--_--.,,,I.-_,
- ! I--I-.L:.I.,,.i.,;,-,.I-I-...I-7.L,:
I
ZI-"-:...I.1,,LI...,._
IrI
-..,,.I,,.,:,,p.-.-I
11,I.
I, -,.:`
I-.I-I:z:,
-,I,-?
,L,-----_.L_I.
:, ,:--.,-,--I,-,:,
.------: I,-,,,.-,:.
,:
.I---.
..I..
'-.IL.--,,,,,I.I:,:-..
;,--,II.: ,-,---LL I.-I.
-::-:
:--,,,-.,,,-,:..---1,:!..,.I-:'
,
-:
- ,-,----, -,-,
,..
,7.,---,,
-t
,Z.
.o
,
,i
I
.
I,I
-.
.',-----.
.:
,-'--Ir..I-,,-:
:
.II.L. L.-I.-III.I.--.-ILI
I-1.-L,,,,,-.
., .--.
I
..
I .,4I..
:,-:.,-...I,-.,.1
.
-:,I-.,.,
,-I.rIr.I
-,.L:;%_
-- ,:'-,I.--...II--I
I4
i,:
I II--:
;:
I..I.,
-I,
:, Li
, :,-,.,-kL :,--.
II..,
-.
L-L-.I-,Ir-,::L,
.. 'II,-i,I....I...I,__.
-.,:-Z.Ir..
6II_,,
,,-.-"11
i.,,
..I-711..1%
- ,-,,'_-:
, ,,,'x
...--,._.II-I-._..__:..
-_:,. ,:,-.I,
f,-II.I-..-,-.---1._,,;
L,.-.,
" ,__M" L,_'-.
.I ,-.,
. .I...I-,IL..
,.1,_-I-.
I.-I.-L
II
...-,
,,.
,_LL.-I--I.-.
..L.;, ._I,,,:, _L
-,-,.q
.
.
1,
--.
,
,
-,
II
,!,oI-...
-1
4:,,
-,:.,:'
,,r,
:,
-,, :,Z,;-,,I-.,
L..,,,
;-I-I.I.
.I'--iiI
.II.-I.L-.
..r.,-I.11.,.-.:
,---,-
,----,I-.I-ff.,--,-.1
-.--0 -,-Ipm---Z-,
,,---- ,iL^...
ll"I..
, Am,
"
."
-,,-`r7-."-.7
,-:,.
'',-I..
,.
-:.--. _,, I.,,...--- -- -.1-.,-...
-,.
---,.,,,;-,"I-rl:
"-:,.-.-
I.r%
I,-,- ,.- --- .-.I1:
:-96-P.-,
I--.-Ki-.
:.
-1I
--- ,-, 1,I, ,-11I-- ,---,
,-.,
,:-,,
I-_,
I ar,''M : , I.
. 1.-I11.
I- il.-, ,,--.1.-:_,,,,,..-,
1,_::
,--.
,...-:-.
-'I,-
, %O-3,,%:-.
INJ
- I Ft.-.---I.
,-., -.'.---
7-.I.-kL
&-;:,.,'.I-.-1:,::--
.- ,
-'
-.-:jt-I.!IL-.
.-,.II.-I.I-I.
-.
--,:--
-,-,,..I.,_--'---,':
. 1.II.
,"..,.
... ,.;
OFTECHOLOG
,,;.........,...-
.............
-
X
'0'y0Et~fV0JEV' H
R)
-00
........
ON EVALUATING EVALUATIONS
by
Richard C. Larson
Leni S. Berliner
OR 096-79
December 1979
Prepared under Grant Number 78NI-AX-0007 from the National Institute
of Law Enforcement and Criminal Justice, Law Enforcement Assistance
Administration, U.S. Department of Justice. Points of view or opinions
stated in this document are those of the author and do not necessarily
represent the official position or policies of the U.S. Department of
Justice.
Table of Contents
Page
Abstract
i
Introduction
1
Evaluation Inputs, Process, and Outcomes
3
Evaluation Inputs
5
Evaluation Process
7
Evaluation Outcomes
12
Why Evaluate Evaluations?
16
Evaluation Documentation Requirements
19
Summary and Suggested Research
21
Notes
23
References
24
Figure 1.
Evaluation Viewed as a System
4
Table 1.
Inputs to an Evaluation
6
Table 2.
Process Components of an Evaluation
8
Table 3
Evaluation Outcomes
15
i
ABSTRACT
Researchers in the field of public program evaluation have recently
been arguing for comprehensive evaluation of programs, evaluations that
examine program inputs, processes, and outcomes.
In this paper we propose
that the same approach is useful for evaluating evaluations themselves.
We identify tentative sets of evaluation inputs, processes, and outcomes
and describe their utility in evaluating evaluations.
Since evaluations
are commissioned to produce information useful to decision makers, it is
suggested that the only fundamental outcomes of an evaluation are
decisions influences by the evaluation.
This fact precludes evaluating
an evaluation solely from the final report of the evaluation, which is
usually written before the relevant decisions are made.
However,
evaluators can and should document useful information about evaluation
inputs and processes, and recommendations along these lines are included
in this paper.
INTRODUCTION
Increased attention has been focused in recent years on the merit
of various public programs.
Legislators, citizens' groups, and others
have been concerned with the extent to which programs meet their
objectives.
programs.
Program personnel want to know how to improve on-going
Researchers as well as legislators have been interested in
programs as social experiments, in attempts to validate or refute various
intervention theories which may later lead to new standardized (nonexperimental) programs.
All of this concern has pushed the burgeoning
field of public program evaluation into the limelight.
many promises:
Evaluation offers
accountability for public expenditures; improvement for
on-going programs; identification of promising new theories leading to
new programs; even generation of new research knowledge.
Yet evaluation to date has seemed to flounder.
evaluation work has been spotty.
ingly small impact.
The quality of
Evaluation findings have had a disappoint-
While considerable attention has been given to the
evaluation of programs, it is somewhat ironic that relatively little
attention has been devoted to evaluating the evaluations (and the
evaluators).
(Except Bernstein and Freeman, 1975; Cook and Gruder, 1978;
GAO, 1978a and b; MSR, 1973; Cook, 1978; Stufflebeam, 1974).
can an evaluation be said to be successful?
On what basis
Attempts to answer this
question have often focused on one sub-element of the evaluation, namely
the technical aspects of evaluation process--as reflected, say, by
-2-
the statistical methodology employed or the survey research methods
used.
This limited focus creates problems analogous to those which
would occur if one conducted an evaluation of a program, exclusively
examining some part of program process, while ignoring program inputs,
outcomes and other aspects of process.
Thus, in the same sense that
evaluation of programs requires a comprehensive analysis of program
inputs, process and outcome, so too evaluation of evaluations requires
analysis of evaluation inputs, process, and outcome.
One of the goals
of this paper is to lay out the elements of such an approach.
In identifying the various components of an evaluation, particularly
outcomes, we are guided by the fact that evaluations are commissioned by
an individual or group to produce information which hopefully has some
decision-consequential impact.
More succinctly,
1.
Evaluation is a process that produces information;
2.
Information is useful only to the extent that it informs
decisions;
3.
A decision is an irrevocable allocation of resources;l
4.
Thus, evaluation is a process that produces information
to assist in the allocation of resources.
The allocation of resources referred to here must be considered broadly.
It can range from obvious programmatic changes (e.g. shifting staffing
patterns or operating procedures), to a change in program budget level
as determined by the funding agency, to a congressional authorization to
spend more (or less) money on similar programs nationally, to a change
in research efforts by one or more researchers who may want to devote
more (or less) of their resources to that programmatic area.
It can
-3-
even refer to decisions by memebers of a program's client group to increase
(or decrease) participation (e.g. allocation of their time and energies)
in the program.
Much confusion, we believe, in the evaluation literature
can be traced to a lack of attention to the identification of decision
makers and decisions to be made (resources to be allocated) as a
consequence of an evaluation.
Our discussion of evaluation outcomes
will be motivated by this decision-oriented point of view.
EVALUATION INPUTS, PROCESS, AND OUTCOMES
Any evaluation is a process, having inputs and yielding outcomes
as illustrated in Figure 1. It is our contention that any comprehensive
evaluation of an evaluation should examine all three of these evaluation
components.
An examination of process alone may, for instance, verify
exemplary technique, but reveals nothing about evaluation impact.
A
review of evaluation outcomes alone is not sufficient for explaining
the causal mechanisms linking evaluation input through process to those
outcomes; indeed, one who analyzes only outcomes is often hard-pressed
to attribute outcomes to the evaluation.
An examination of inputs alone
reveals little more than the collection of resources mustered to conduct
the evaluation.
In this section we present an initial listing of detailed
elements comprising evaluation input, process, and outcome.
-4-
Evaluation Inputs
EVALUATION
Evaluation Outcomes
.
PROCESS
Figure 1:
Evaluation Viewed as a System
r1
-5-
Evaluation Inputs
The inputs to an evaluation may be considered to be an inventory
of resources and methodologies brought to bear on the evaluation, and
a description of the evaluation setting.
is summarized in Table 1.
Our proposed set of inputs
Obvious inputs include evaluation budget
(both in absolute terms and as a percentage of program budget), duration
of the evaluation, timing of the evaluation with respect to the program
being evaluated, and skills (and other attributes) of the evaluation
personnel.
Despite the indisputable importance of these items and
despite the urgings of evaluators to consider program inputs during
evaluations, few evaluators themselves document these rudimentary
evaluation inputs.
Any comprehensive evaluation of an evaluation is
thwarted at an early stage if these inputs are not known;
Other necessary inputs also listed in Table 1 include attributes
of the program being evaluated and its personnel (e.g. training and
experience, determining their attitudes toward the evaluation),
evaluation methodology and design, and audience or client group for the
evaluation, and the purpose of the evaluation.
Here, purpose of the
evaluation refers to the resources which one or more client groups
(decision makers) are likely to (re)allocate as a consequence of
information provided by the evaluation.
A description of evaluation
methodology and design should include not only the statistical procedures
to be used to analyze data, but also the entire plan for considering
program inputs, process and outcomes.
Moreover, it should indicate
milestones along the way, points at which information may be fed back
-6-
Table 1
Inputs to an Evaluation
Evaluation inputs; An inventory of resources and methodologies brought
to bear on the evaluation and a description of the evaluation setting.
(i) Budget of the evaluation (and other material resources
available to the evaluators).
(ii) Duration of the evaluation.
(iii) Timing of the evaluation with respect to the program being
evaluated.
(iv) Attributes of evaluation personnel (e.g., training, experience,
"world view").
(v) Attributes of program personnel (e.g., experience, commitment,
education).
(vi) Attributes of the program being evaluated (e.g., goals, substantive
area of concern, clieht group).
(vii) Evaluation methodology and design.
(viii) Audience, or "client group", or decision makers for the evaluation
and purposes of the evaluation.
-7-
to program staff for possible program modification, and other points at
which information obtained so far may be used adaptively to change
evaluation and/or program procedures; rules for such adaptive change,
to the extent possible, should be stated explicitly a priori (i.e.,
such rules for adaptiveness are themselves evaluation inputs).
Evaluation Process
Evaluation inputs are utilized to develop information which leads
to evaluation outcomes (i.e., decisions to be made).
of inputs is evaluation process.
This utilization
Paralleling the idea of program process,
evaluation process is the actual conduct of the evaluation as compared
with that planned in the evaluation design.
evaluation process are given in Table 2.
Our proposed components of
As the discussion will indicate,
this list is meant to be illustrative, but by no means exhaustive.
The
first three elements of evaluation process relate to the interaction
between the evaluators and the program being evaluated.
Obviously
important are (i) the types, intensity, and frequency of interactions
between evaluators and program staff members.
This, in turn may lead
to (ii) (an inadvertent) response of program staff and client groups to
the presence of the evaluators (e.g., "Hawthorne effect").
While this
response may be difficult to detect and measure, it constitutes an important
element of evaluation process, and attempts should be made to estimate and
report it.
The final element of evaluator/program interaction is (iii)
the extent to which information acquired during the evaluation is fed
back to program staff, perhaps modifying program procedures.
This reflects
deliberate transfer of information that may result in changes in program
-8-
Table 2
Process Components of an Evaluation
Evaluation process: Actual conduct of the evaluation compared with
that planned in the evaluation design.
(i)
(ii)
(iii)
(iv)
(v)
(vi)
(vii)
(viii)
Types, intensity and frequency of interactions between
evaluators and program staff members.
Response of program staff and client groups to the presence
of evaluators.
Extent to which information acquired during the evaluation
is fed back to program staff, perhaps modifying program
procedures.
Extent to which information acquired during the evaluation
is used to modify the allocation of evaluation resources.
Adaptivness of evaluation design (i.e., capacity to respond
to changes in the program) and history of adaptations.
Turnovers in personnel (e.g., evaluators, program staff,
client groups of both program and evaluation).
Testing of hypotheses regarding the program.
Documentation of findings.
-9-
operation, in contrast to inadvertent interactions identified in (ii)
which may also affect program operation.
Item (iii)
relates to the
controversy in evaluation about whether or not the evaluators should
"board the train" of the program or ride along in a different train on
parallel tracks (Rossi and Wright, 1977).
On the same train, they can
provide information thought to be useful during the course of the
evaluation, usually yielding changes in the program during the evaluation.
The two-train model precludes such useful exchange of timely information.
In our view, a reasonable way to tackle this question is to view the
choice as being directly dependent on the intended audience or client
of the evaluation.
If, for example, the evaluation is being done
primarily for the purposes of the program clients or program staff, then
it appears that the evaluator(s) should "board the train."
When, however,
the evaluation is being done as part of an independent assessment for a
funding agency, then to avoid possible conflicts of interest the "two
train" approach seems more appropriate.
The next two items relate to the adaptability of the evaluation.
First, one is interested in (iv) the extent to which information acquired
during the evaluation is used to modify the allocation of evaluation
resources.
Is the evaluation in a "straight jacket" design, or do the
evaluators modify the design in response to information obtained so far
in the evaluation?
Adaptability may be reflected in elements of process
evaluation such as the allocation of participant observers and/or
interviewers to various parts of the program.
Or it could relate to the
sequential adaptive generation and testing of alternative hypotheses
-10-
regarding program operation.
It is our conjecture that many evaluations
in practice are adaptive, but, lacking rules and encouragement for
adaptability, the evaluators in their reports are likely not to describe
this element of their evaluation.
A related issue is (v) the adaptive-
ness of the evaluation design in response to changes in the program being
evaluated.
For instance, during the operation of the program, an
employee strike could occur, a new relevant law could be enacted, or a
citizens' group could protest against some particular aspect of the
program.
To what extent is the evaluation jeopardized by such program
changes and interruptions, and to what extent can it adapt to them?
This concern stands in contrast to (iv), in which the program was
assumed fixed, but evaluative information about components of the program
changed over time, perhaps resulting in changes in evaluation tactics.
No evaluation design can stand impervious to all conceivable unforseen
changes in the program and its operating environment, but some are more
robust than others.
A chronological history of adaptions of the evaluation
to changes in the program would seem to be an important part of evaluating
evaluation process.
The next item is (i)
turnover in personnel (e.g., evaluators,
program staff, client groups of both the program and the evaluation).
This is one of perhaps several internal unplanned changes in program or
evaluation process.
However, it appears to be a critical one, in that a
change in one or more key persons in the evaluation or in the program
can markedly affect the outcome of both.
We are aware, for instance,
of at least one evaluation that had three different directors during
the course of its operation; one can speculate as to the quality and
-11-
impact of the final evaluation.
A change in the client group of either
the program or the evaluation is also important.
For instance, a
significant fraction of evaluations that have had little or no
eventual
decision impact appear to fall victim to the "vanishing advocate"
syndrome, in which the person who originally commissioned the evaluation
has moved to another professional position, only to be replaced by someone
unsympathetic to the original purposes of the evaluation (Chaiken et al.,
1975).
The seventh entry in Table 2, (vii) testing hypotheses regarding
the program, is one of perhaps many elements of analysis of evaluation
findings.
We believe that all such analyses, both statistically and
subjectively based, constitute a part of evaluation process, not outcome.
It appears to be this element of evaluation process--evaluation technique
and methodology as applied in practice--that has received most scrutiny
by evaluators of evaluations.
Perhaps this is because manipulation of
numbers is one of the few elements of evaluation process that can be
replicated and scrutinized by others after termination of the evaluation.
And statistical procedure is one of the few components of evaluation
process in which apparently universal "scientific" measures of accountability can be applied.
But one should not fall prey to the trap of
misplaced emphasis on statistical method.
A statistically elegant
evaluation may be seriously flawed in other respects; and statistical
correctness by no means guarantees decision impact.
On the other hand,
a statistically flawed evaluation may indeed present imperfect information to decision makers.
The imperfections may lead to decisions that
would have been improved if more accurate information had been available.
-12-
the "costs" of such imperfect information can be considerable.
Yet, when
balanced with other components of evaluation process, it is quite possible
that a statistically flawed evaluation can still present useful information to decision makers--where usefulness implies decisions being made
that are in some sense "better" than those that would have been made in
the total absence of the evaluation.
In order to place in perspective
the importance of statistical procedure and to estimate the cost of
statistical error, we would argue strongly for a comprehensive evaluation
of evaluation process, as reflected by the other elements in Table 2.
The eighth and final entry in the Table is (viii) documentation of
findings, i.e., creation of the "final report."
While the final report
may represent a key product of the evaluation, we believe that it is
related to evaluation process; it is not a fundamental evaluation
outcome.
Like several other entries in Table 2, this one also could be
broadened, perhaps to "communication of final evaluation findings."
Certainly the final report--its structure, content, level and style of
presentation--is an important part of the communication process.
But
also important are oral presentations, use of teaching aids to convey
the essential results, and other activities and devices for communication
and dissemination of results.
For instance, a methodologically flawless
evaluation whose findings are unintelligible to decision makers will have
at best marginal impact.
Evaluation Outcomes
For many public programs, the boundary separating process and
outcome and their corresponding measures is indeed fuzzy.
As one
-13-
example, "citizen and client satisfaction" with a particular program has
been considered by evaluators both as a process measure and an outcome
measure.
However, improved citizen satisfaction as an outcome measure is
problematic; its basis may be illusory (e.g., citizens may feel safer
when true crime rates increase; students who "like" their teachers may
in fact be learning less, etc.).
Extending this example to our concern
here, program staff satisfaction with an evaluation may be well founded
if the evaluators--funded by the program in order to help improve the
program--helped the staff people to discover ways of making their
program more effective.
On the other hand, such positive feelings
toward the evaluators may be little more than self-serving if the
evaluators-commissioned by the program's funding agency--find nothing
to recommend discontinuance of the program.
Again, a dilemma in
evaluation is resolved in part by examining the primary client group
of the evaluation.
Any attempt to demarcate the boundary between evaluation process
and outcome and measures of their effects is done in the presence of
ambiguity and controversy.
Still, the inherent difficulties should not
act to preclude discussion on this vital matter.
In an attempt to provide
one input to the debate, we take a relatively firm stand on evaluation
outcome, motivated by our decision orientation:
ultimate outcomes of an
evaluation are the decisions (resource allocations) influenced by the
evaluation.
All other outcomes are process outcomes.
Most evaluators
discover the decision consequences of their evaluation only long after
submission of the final evaluation report, if at all; because of this it
-14-
is inappropriate for those who evaluate evaluations to judge their impact
only from reading the final report (Larson et al., 1979).
The time period
of the evaluation of an evaluation must extend beyond that of the
original evaluation in order to assess its ultimate decision consequences.
As a first cut, we have identified the different types of decisions
that may be affected by an evaluation as follows:
(i)
"the funding agency's decision": decision by funding agency
to fund, refund, modify or cancel the program;
(ii)
"the program staff's operational decision":
decision by
program staff to modify any of the program procedures;
(iii)
"the program client group's decision":
decision by members
of the program client group to increase, decrease or
otherwise alter participation in the program (assuming
they are informed of the evaluation to begin with);
(iv)
"the research community's decision":
decision by one or
more members of the research community to study (or not to
study) further the questions raised in the evaluation;
(v)
"the decision of those involved in related programs":
decision by one or more other funders and/or program personnel
in other jurisdictions to initiate, modify or terminate
similar programs.
Decision makers under (iv) and (v) above may be said to be "second-order
decision makers," since they are not directly involved with the program
being evaluated or the evaluation.
The evaluation's ultimate impact on
these decision makers will in many cases be very difficult to assess.
-15-
TABLE 3
Evaluation Outcomes
Evaluation Outcomes:
A listing of decisions influenced by the evaluation.
(i) Decision-by funding agency to fund, refund, modify
or cancel program.
(ii) Decision by program staff to modify any of the program
procedures.
(iii)
Decision by members of the client group to alter
participation patterns in the program.
(iv) Decision by one or more members of the research community
to study further the questions/issues raised in the evaluation.
(v) Decision by one or more other funders and/or program
personnel (in other jurisdictions) to initiate, modify,
or terminate similar programs.
-16-
Recalling that retention of the status quo is also a decision (although
occasionally it is mislabeled as "no decision"), then it is useful to
note that each of the decisions cited above will be made whether or not
the evaluation is conducted.
(For convenience, the decisions above are
summarized in Table 3.)
WHY EVALUATE EVALUATIONS?
Evaluations are performed to provide information about a program to
decision makers.
Thus, evaluations of evaluations are performed to provide
information about an evaluation to a possibly different set of decision
makers.
For instance, an evaluation of an evaluation can provide an
independent assessment for decision makers of the quality of the information
presented in the evaluation.
This would enhance the extent of "informedness"
of the resulting decisions, but at a cost--the cost of the evaluation of
the evaluation.
Clearly if this cost exceeds some threshold, its marginal
information value may not be adequate to justify its cost.
In this context,
the decision to evaluate an evaluation is also an allocation of resources
which may or may not be justified at a particular point in time, given
one's knowledge about the original evaluation, the program being evaluated,
and the marginal cost and expected marginal information content of the
evaluation of the evaluation.
In another application, a collection of evaluations of evaluations
(and evaluators) could be used in the selection of an evaluator of a
-17-
program.
The selection criterion would probably include a trade-off
between evaluation quality and cost.
Clearly other reasons for evaluating evaluations abound.
One,
for instance, is to pool together in a "meta evaluation" framework
the results of many different evaluations of similar programs to obtain
synthesized research results.
This and other examples of evaluating
evaluations are discussed by Cook and Gruder (1978).
One perhaps not so apparent reason for evaluating evaluations relates
to the evaluation enterprise itself.
Suppose one or more sets of decision
makers are concerned with examining evaluation per se, and are not
concerned directly with the program being evaluated.
For instance,
a potential funding agency may be interested in evaluating a particular
evaluator, or a researcher may be interested in evaluating a new
experimental design procedure.
Then, in what we have discussed so far,
"program" becomes "evaluation," i.e., that set of activities having
programmatic goals and objectives.
to evaluating a program.
Evaluating the evaluation is analogous
Correspondently, there is a set of evaluation
inputs, processes, and outcomes directly analogous to those listed in
Tables 1, 2 and 3 for the evaluation of a program.
As an example, it is
instructive to consider the possible decisions that could be affected by
information produced by the evaluation of an evaluation.
The "funding
agency's decision" could relate to supporting additional research related
to the specific methodology employed in the (original) evaluation. 3
The
"program staff's operational decision" could refer to potential modifications of evaluation conduct by the (original) evaluators.
The "program
client group's decision" could relate to decisions by those awaiting
-18-
information from the (original) evaluation to utilize the information
in different ways.
The "research community's decision" would most likely
be a decision by one or more members of the research community to study
(or not to study) further issues related to that type of evaluation.
The "decision of those involved in related programs" could be the
decision by one or more evaluation funders and/or evaluation personnel
in other areas to initiate, modify, or terminate similar types of
evaluations.
Thus, the comprehensive evaluation of evaluations has the
potential to provide the full spectrum of information and decision
consequences for evaluations that is expected of evaluations of programs
for programs.
One final note of the utility of evaluating evaluations:
any
evaluator of evaluations--like any other evaluator--is imperfect, requiring
then an evaluator to evaluate evaluators of evaluations.
But this new
third level evaluator is also imperfect, requiring yet another evaluator.
This potentially infinite morass can be avoided by considering the
expected marginal costs and informational utility of the concept of the
expected value of perfect information; this is the expected improvement
in the decision maker's outcome measure that can be attributed to obtaining
perfect information (i.e., eliminating uncertainties) at the decision stage.
No decision maker should support a hierarchical evaluation superstructure
having an expected cost greater than the expected value of perfect
information (Thompson, 1975).
-19-
EVALUATION DOCUMENTATION REQUIREMENTS
The framework described in the previous sections has implications
for documentation of program evaluations.
Given our definition of
comprehensive evaluation and our listing of inputs, processes and
outcomes, the information required to do a comprehensive evaluation of
evaluations can become considerable.
Whether or not an independent
evaluation of an evaluation is to be performed, documentation in the
final evaluation report of inputs, processes and outcomes (as far as
possible) of the evaluation would seem to be essential.
As we discovered in an empirical study of criminal justice program
evaluations (Larson et al., 1979), current evaluation documentation
practice is uneven and sorely lacking.
Of the roughly 200 studies
in the sample, only 4 percent indicated the percentage of the program
budget allocated for the evaluation, and only 2 percent indicated total
evaluation budget.
Thirty-one percent of the reports in the sample
indicated the total duration of the evaluation, while 8 percent at least
indicated (though not aways expZlicitZy) the timing of the evaluation
with respect to the program being evaluated.
None of the reports
described the professional or other attributes of either program staff
or evaluation personnel.
Finally, while 90 percent of the reports made
at least some reference to the context or purpose of the effort, actual
potential users of the evaluations were rarely identified explicitly.
Only 58 percent of the reports contained an analysis of program goals,
and only 47 percent discussed the program's client group in any way.
-20-
Evaluation process components fared much worse than evaluation
inputs in the sample of final reports.
The constituent parts of
evaluation process listed in Table 2 were rarely if ever included in
the final reports.
In our overall review of the evaluation research
literature, we have found that elements of evaluation process do appear
in the growing number of anecdotal reports on non-utilization (Weiss,
1977).
But there appears to be little tradition of evaluators routinely
reporting on their own evaluation process.
Such lack of self-reporting
reduces the ability of decision makers to assess the quality of information
produced by the evaluation.
And, information on evaluation process could
only help to enhance evaluators' and program managers' awareness of
evaluation limitations and pitfalls, thus leading to improvement of
evaluation practice.
Evaluation outcomes in terms of decisions influenced by the evaluation
are rarely documented in the final report, due in part to the timing of
the final report with respect to decisions yet to be made.
Even after
decisions are made, it is often exceedingly difficult to estimate what
influence (if any) the evaluation had on the decisions.
Here, it seems,
we need new methods for follow-up attribution and documentation.
Thus, at the current time, we can only make a plea for more complete
self-reporting of evaluation inputs and process.
Self-reporting is open
to criticism on grounds of objectivity, particularly in the area of
evaluation process.
Yet even imperfect information in this area would
be more valuable than the present state of nearly no information.
Particularly for those second-order decision makers not directly
-21-
affiliated with the program being evaluated, it seems that at least
rudimentary knowledge of evaluation inputs and process would be
necessary to assess the possible relevance of the findings to them.
SUMMARY AND SUGGESTED RESEARCH
We have proposed that evaluation is a process producing information
that can be evaluated on the basis of its relevance to decision makers.
Paralleling program operation, an evaluation too can be characterized by
inputs, process, and outcomes.
We have provided a suggested list of
each, arguing that the only ultimate outcomes of an evaluation are
decisions influenced by the evaluation.
Thus, however difficult to measure,
the impact of an evaluation must be judged on the basis of resources
(re)allocated as a consequence of evaluation information provided.
of evaluations per se have several potential purposes:
Evaluations
to provide an
independent assessment to decision makers of the quality of information
contained in an evaluation; to provide guidance in selecting an evaluator;
to assimilate "research knowledge" from a number of separate but similar
programs; to provide a vehicle for examining the evaluation enterprise
itself.
Our concern for evaluation inputs, processes and outcomes
extends to recommendations for improved evaluation documentation in these
areas.
Evaluating an evaluation utilizing only inputs, or only process,
-22-
or only outcomes suffers drawbacks identical to similarly limited
evaluations of programs.
The need for comprehensiveness is apparent,
but the means for bringing it about in practice is not.
Thus, further work is needed in attempting to devise methods for
carrying out comprehensive evaluations of evaluations, within time and
budget constraints that are acceptable to potential decision makers.
Numerous important questions abound:
evaluations?
Who should conduct evaluations of
When is self-reporting of evaluation inputs and process
adequate?
How do we measure the effect of evaluation information on a
decision?
How do we historically recreate a decision maker's state of
(imperfect) knowledge at the time of-decision?
If a follow-on program
is instituted to measure an evaluation's ultimate impact, who should
fund it and who should do it?
Each of these questions, and more, should
provide fruitful areas for future research.
-23-
NOTES
1.
This definition of decision is taken from Howard (1966).
"'Irrevocable' does not imply 'for all time', but at least for the next
short time interval that an allocation of at least one resource has
been made.
That is, a decision is not a 'decision to.make a decision',
but rather the concrete action implied by the decision.
After any time
interval, a decision may be replaced by another decision, perhaps based
on updated information."
2.
The small Roman numeral in parentheses identifies the particular
point in Table 2.
3.
We use the term "(original) evaluation" to indicate the evaluation
being evaluated.
-24-
REFERENCES
BERNSTEIN, I.N. and H.E. FREEMAN (1975) Academic and Entrepreneurial
Research:
Srudies.
The consequences of Diversity in Federal Evaluation
New York:
Russell Sage Foundation.
CHAIKEN, J., L. CRANBILL, L. HOLLIDAY, D. JAQUETTE, M. LAWLESS, and
E. QUADE (1975) Criminal Justice Models:
Monica, CA:
An Overview.
Santa
Rand Corporation Technical Report R-1859-DOJ.
COOK, T.D. (1978) "Utilization, Knowledge-Building, and Institutionalization:
Three Criteria by Which Evaluation Research Can be Evaluated,"
(Introduction) in T.D. Cook (ed.) Evaluation Studies Review Annual,
Vol. 3.
Beverly Hills:
Sage.
COOK, T.D. and C.L. GRUDER (1978) "Metaevaluation Research." Evaluation
Quarterly 2(February):
5-51.
HOWARD, R.A. (1966) "Decision Analysis:
Applied Decision Theory," in
D.B. Hertz and J. Melese (eds.) Proceedings of the Fourth International
Conference on Operations Research. New York:
Wiley.
LARSON, R.C., V.M. BIER, E.H. KAPLAN, C. MATTINGLY, T.J. ECKELS,
N. REICHMAN, and L.S. BERLINER (1979) Interim Analysis of 200
Evaluations on Criminal Justice. (LEAA Grant 78NI-AX-0007).
Operations Research Center.
Institute of Technology.
Cambridge, MA:
Massachusetts
-25-
MINNESOTA SYSTEMS RESEARCH, INC. (1973) Proposed Review Systems for
the Evaluation of Technical Quality & R & D Projects in the
Department of Health Education and Welfare.
ROSSI, P.H. and S.R. WRIGHT (1977) "An Assessment-of Theory, Practice,
and Politics."
Evaluation Research l(February):
STUFFLEBEAM, D.L. (1974) Meta-Evaluation.
5-52.
Michigan University Evaluation
Center, Occasional Paper #3.
THOMPSON, M. (1975) Evaluation for Decisions in Social Programmes.
England:
Westmead,
Saxon House, D.C. Heath Ltd.
UNITED STATES GENERAL ACCOUNTING OFFICE (1978a) Status and Issues:
Program Evaluation. (PAD-78-83).
Washington, D.C.
(1978b) Assessing Social Program Impact Evaluations:
A Checklist
Approach (Exposure Draft) (PAD-79-2), Washington, D.C.
WEISS, C.H. (1977) Uising Social Research in Public Policy Making.
Lexington, MA:
Lexington Books.
Federal
Download