So What? Using Outcome-Based Evaluation to Assess the Impact of

advertisement
So What? Using Outcome-Based Evaluation to Assess the Impact of Library Services
So What? Using Outcome-Based
Evaluation to Assess the Impact of
Library Services
Rhea Joyce Rubin
for the
Massachusetts Board of Library
Commissioners
June 2004
Rubin for the Massachusetts Board of Library Commissioners_2004____________________________1__
So What? Using Outcome-Based Evaluation to Assess the Impact of Library Services
OUTCOME-BASED EVALUATION WORKSHOP
AGENDA
9:30 – 10:00
Registration and Coffee
10:00 – 12:00
Why Use Outcome-Based Evaluation?
Concepts and Terminology
Selecting Outcomes
12:00 – 12:30
Lunch Break
12:30 – 3:00
Selecting Indicators and Targets
Writing an Outcome-Based Evaluation Plan
3:00
Adjourn
Rubin for the Massachusetts Board of Library Commissioners_2004____________________________2__
So What? Using Outcome-Based Evaluation to Assess the Impact of Library Services
KEYS TO UNDERSTANDING
OUTCOME-BASED EVALUATION
Outcome = Impact on End User
Impact =





Changes in
Behavior
Attitude
Skills
Knowledge
Condition/ Status
Outcome-based evaluation = A user-centered approach to assessment of
programs/ services that are provided to address particular user needs
and designed to achieve change for the user.
From the user in the life of the library
to the library in the life of the user.
Rubin for the Massachusetts Board of Library Commissioners_2004____________________________3__
So What? Using Outcome-Based Evaluation to Assess the Impact of Library Services
THE VALUE OF OUTCOME- BASED EVALUATION

Useful as a planning tool; requires needs assessment

Clarifies the purpose of the program/ service

Keeps staff and stakeholders focused on goals by stressing “so what”
rather than process

Stimulates discussion of issues

Helps keep implementation on track; milestones are identified

Indicates when changes to program are needed

Energizes staff by demonstrating the real, human impact their work
produces and by stressing common purposes and goals

Gives insight into why and how library services are used; new
perspective on library services in context of the user

Assists in fundraising & grant writing by providing statistics on results

Provides empirical evidence that what you are doing is what you
intended to do

Quantifies the anecdotes & success stories

Identifies effective programs/services

Demonstrates the library’s contribution to solving community problems

Demonstrates accountability as required by IMLS under the
Government Results and Performance Act of 1993.
Rubin for the Massachusetts Board of Library Commissioners_2004____________________________4__
So What? Using Outcome-Based Evaluation to Assess the Impact of Library Services
INTRODUCTION TO TERMINOLOGY
Inputs:






Are resources devoted to or used by a program
Collections, materials, equipment and supplies
Staff and volunteer time and expertise, community partners
Needs assessment findings and other background information
Facilities, computers, online access
Also, constraints on programs such as laws, regulations, funding
requirements
Outputs:





Answer “How many?” (extensiveness)
Are measures of product volume (i.e. number of services or products
provided) or evidence of service delivery (i.e. number of participants)
Are the results of inputs (resources) and activities (programs or
services)
Are from the staff perspective
Are objectively quantified by neutral observers
Outcomes:






Answer “So what?” (effectiveness)
Are measures of impact or benefit to end user, usually reported in
amount of change in skills, knowledge, attitude, behavior, or condition
Are also the results of inputs (resources) and activities (programs or
services)
Are from the participant/ customer’s perspective
Are often quantified by participants’ or others’ perceptions (i.e. selfreports or anecdotal evidence)
May be measured through professional assessments (e.g. testing)
Rubin for the Massachusetts Board of Library Commissioners_2004____________________________5__
So What? Using Outcome-Based Evaluation to Assess the Impact of Library Services




Are the success stories of outputs
Are best used in conjunction with output measurement
Make a project’s expectations explicit
Present assumptions of cause and effect, not concrete scientific
evidence
Rubin for the Massachusetts Board of Library Commissioners_2004____________________________6__
So What? Using Outcome-Based Evaluation to Assess the Impact of Library Services
OUTCOME-BASED EVALUATION GLOSSARY
Note: Terms are listed in alphabetical order.
Activities Programs or services (such as tutoring, lectures, storytelling)
provided to the end user by the library. The specific jobs (such as recruiting
and training volunteers, publicizing a lecture series) are called tasks. Answers
the question: “What does the library do?”
Candidate Outcomes are potential outcomes that must be verified. Usually
the result of brainstorming and assumptions by library staff members,
candidate outcomes must be checked with participants and others before
being accepted as the correct measures of a program. (See “Selecting
Outcomes to Measure.”)
Community Goals relate to the needs of the community and are beyond the
scope of the library alone; the library’s program is usually one part of a larger
effort to meet a community goal. (For example, increase school readiness
among elementary school children, decrease juvenile delinquency in teens,
increase socialization by isolated elders.) Answers the question: “What does
the community hope to achieve in response to an identified community
need? (Note- you are not being asked to define community goals, but when
you document why a project is needed, it should make sense of terms of how
the library is relevant to a community achieving such a goal. )
Goals In evaluation, there are two types of goals. Both are developed in
response to a demonstrated need, broad in scope and long range. In
traditional evaluation, goals reflect the mission of the library. In outcomebased evaluation, goals reflect the ultimate impacts desired for the targeted
users, and usually relate to both larger community goals and the library’s
mission statement. Provides a compass for the program. Answers the
question: “What does the library hope to achieve for the users?”
Rubin for the Massachusetts Board of Library Commissioners_2004____________________________7__
So What? Using Outcome-Based Evaluation to Assess the Impact of Library Services
Immediate outcomes See “Intermediate outcomes.”
Indicators Measurable characteristic or change which indicates achievement
of the outcome. Behaviors which demonstrate changes in knowledge, skills, or
attitudes, or which imply such changes. Amount of expected change is part of
the indicator. (E.g. increase in one reading level within one year; ability to use
a database independently; successful completion of sample citizenship exam).
To gauge success for the participant. Answers the questions: “How will we
know that participants experienced the intended outcome(s)?” and “What
will we measure?”
Initial outcomes See “Intermediate Outcomes.”
Inputs Resources (such as money, staff, volunteers, facilities, collections,
community partners) used to plan and provide a program or service. Also
constraints on a program such as laws, regulations, funding requirements.
Answers the question: “What do we have to work with?”
Interim outcomes See “Intermediate Outcomes.”
Intermediate outcomes Milestones in the life of a project, events which are
necessary for successful outcomes. Points at which project staff must decide
whether to continue current activities or to modify them in order to achieve
the desired outcomes. (E.g. In a family literacy project with the desired
outcome of increasing family reading at home, attendance at library story
times may be considered necessary. So the number of participants signing up
for the first story time may be a milestone in the project.) Often users
satisfaction is an intermediate outcome because it is necessary for users to
be satisfied with a service in order for them to use it repeatedly. Sometimes
called initial, immediate, short-term or interim outcomes. Answers the
questions: “What is the short-term benefit to the user from this
program/service?” and :What will the user do that is necessary if s/he is
to achieve the long-range outcome?
Logic Model See “Outcome-Based Evaluation Plan.”
Rubin for the Massachusetts Board of Library Commissioners_2004____________________________8__
So What? Using Outcome-Based Evaluation to Assess the Impact of Library Services
Measurement Verification of outcomes using the indicators in relation to the
selected targets. Data collection methods include interviews, pre and posttesting, professional observation, self-administered surveys.
Outcome-Based Evaluation Plan Statement of library goal; description of
whom the program expects to affect; inputs; services/programs; outputs;
outcomes; indicators; targets; and data collection plan for the entire program.
Sometimes called a logic model.
Outcome Measure Statement Assertion of indicator + target + data
collection method for a specific outcome. E.g. “90% of program participants
will double the amount of time reading to their children as measured on a preand post- program questionnaire.”
Outcomes Benefits or impacts to the end user which demonstrate the
effectiveness of the program or service. Usually are changes in knowledge,
skills, attitude, behavior, or condition. Answers the question: “What
difference did our program make to the participant?”
Outputs Products resulting from the inputs and activities of the library.
Volume of successful activities or evidence of service delivery. Objectively
quantified measures (such as number of books circulated, number of hours
devoted to homework assistance, or number of attendees at an author
reading). In traditional evaluation, projected outputs with a timeline
attached are referred to as objectives. Answers the question: “How much
did the library do?
Short-term outcomes See “Intermediate outcomes.”
Targets Numerical standards or criteria of success for your program. The
expected proportion of outcomes. (E.g. 75% of participants reporting a
specified effect six months after a program; 20% of attendees at resume
writing workshops report getting job interviews). Some outcomes are defined
qualitatively and do not have a specified target. To gauge success of the
library’s program. Answers the question: “How will we know that our
program is a success?”
Rubin for the Massachusetts Board of Library Commissioners_2004____________________________9__
So What? Using Outcome-Based Evaluation to Assess the Impact of Library Services
Tasks The specific jobs (such as recruiting and training volunteers or
publicizing a lecture series) staff or volunteers must do to be able to provide
successful programs or services. Key steps needed to provide or manage
services.
Rubin for the Massachusetts Board of Library Commissioners_2004____________________________10__
So What? Using Outcome-Based Evaluation to Assess the Impact of Library Services
GLOSSARY EXEMPLIFIED:
Information Literacy Program in a Public Library
Organizational Mission: [unique to each grantee]
Program Purpose: Train students to use electronic information competently
and independently.
Inputs: Bibliographic instruction librarian, contractual consultants,
collaboration with school librarians, computers, software, electronic
resources, print and non-print materials, computer lab, grant funds and local
match.
Program Services: Classes, ongoing support and mentoring, web tutorial.
Outputs: 100 flyers designed, produced, and distributed at 6 schools. Course
materials developed. 10 information literacy classes held, each of which is
attended by 10 students.
Immediate (Initial) Outcome (Milestone): Students can differentiate among
three types of electronic resources (licensed databases, OPACs, and freely
available web pages).
Indicator of Immediate Outcome: Students are able to correctly distinguish
among three kinds of resources, with at least 75% accuracy.
Target for Immediate Outcome: 80 of 100 (80%) of students who complete
the program will demonstrate the indicator.
Intermediate Outcome (Milestone): Students know how to find reliable,
objective, accurate and current information in different electronic resources.
Indicator of Intermediate Outcome: Students demonstrate their ability to
search an OPAC, the WWW, and licensed databases to find a specified piece
of information, with at least 75% accuracy.
Rubin for the Massachusetts Board of Library Commissioners_2004____________________________11__
So What? Using Outcome-Based Evaluation to Assess the Impact of Library Services
Target for Intermediate Outcome: 80 of 100 (80%) of students who
complete the program will demonstrate the indicator.
Long-range Outcome: Students meet a personal or educational goal using
information they found in electronic resources.
Indicator of Long-range Outcome: Students report that they have met a
personal or educational goal using accurate and current information they
found in electronic resources.
Target for Long-range Outcome: 70 of 100 (70%) of students who complete
the program will demonstrate the indicator.
Rubin for the Massachusetts Board of Library Commissioners_2004____________________________12__
So What? Using Outcome-Based Evaluation to Assess the Impact of Library Services
CHARACTERISTICS OF PROGRAMS/SERVICES
BEST SUITED TO OBE
 Designed in response to a demonstrated need
 Contribute to a larger community goal and to the library’s mission
 Aim to impact people rather than produce or deliver a product
 Focus on effectiveness, not efficiency
 Concentrate on results, not process
 Has a distinct beginning and end
 Has stable program staff (including volunteers)
 Has supportive administration and stakeholders
 Have voluntary participation by users
Intend to change behavior, attitude, skill,
knowledge, or condition of a specific
group of people in response to a need.
Rubin for the Massachusetts Board of Library Commissioners_2004____________________________13__
So What? Using Outcome-Based Evaluation to Assess the Impact of Library Services
SELECTING OUTCOMES TO MEASURE

Identify your intended participants first.

Gather ideas on positive (and negative) outcomes for your program.
 Brainstorm candidate outcomes with staff. Ask “Who benefits from
this service/ program and how?” Then, “In the best of
circumstances, what other benefits might there be for those
participants?” Note that talking to staff will not yield all the
outcomes and is only a first easy step.
 Interview or hold focus groups with current and past participants.
Ask: “If this program really helps you, how will you be better off?”
or “What do you expect to change for you as a result of this
program?””
 Talk with staff and volunteers who work directly with your intended
participants at other programs at your library or elsewhere. Ask:
“What changes might we see in participants if our program is
successful?”
 Talk with representatives of agencies that might be the “next step”
for former participants. Ask: “What will participants need to know
or be able to do in order to succeed in your program?”
 Review agency or program records to discover what outcomes have
been noted informally. Ask staff and others for anecdotes and
testimonials they have heard.
 Ask about unintended negative outcomes so you can avoid them (e.g.
cheating on summer reading records in order to get prizes).
Rubin for the Massachusetts Board of Library Commissioners_2004____________________________14__
So What? Using Outcome-Based Evaluation to Assess the Impact of Library Services

If you already have a program idea, test it by brainstorming the “if-then”
chain of influences and benefits. If the program does such and such, we
can assume such and such change in the participants. For example, if the
proposed library program is homework assistance, think through these
steps:
 If students have assistance with their homework  then they will
do better at their school work. If they do better at school  then
they will get better grades and attend school more regularly.
 If they get better grades and attend school more regularly  then
they are more likely to graduate. If they graduate  then they are
more likely to become employed.

If you do not yet have a program, but want to develop one to meet a given
goal, the “if-then” chain will work from intended results back to potential
services. To use the same example, if your community goal is higher
employment rate among young adults, consider:
 If we want more young adults to be employed  then they must
be prepared with a high school education. If we want more young
adults to graduate high school  then we need them to attend
school more regularly and to get better grades.
 If we want them to do better in school  then we need to offer
assistance with their homework.

Be sure that the If and the Then phrases relate to each other and are
realistic.

Select outcomes that are valuable for the participants, achievable by the
program, and relate to the library’s priorities.

Remember that some intermediate outcomes may be identical to outputs.
For example, the number of people attending a program or satisfied with a
service can be an output or an intermediate outcome, depending on the
situation.
Rubin for the Massachusetts Board of Library Commissioners_2004____________________________15__
So What? Using Outcome-Based Evaluation to Assess the Impact of Library Services

The more immediate an outcome, the more influence a program has on its
achievement. The longer-term the outcome, the less direct influence a
program has on its achievement and the more likely that other forces
intervene.

Keep in mind that a program’s outcomes must have an appropriate scope.
To clarify if an outcome is reasonable for your program, ask: “Can our
program alone have a significant influence on this outcome?” Of course,
the library cannot take all the credit or blame for a change in participants’
lives.

Outcome-based evaluation is more difficult if the program/ service has
one time or short term contact with an individual; if follow-up is
impossible; if the goal is to prevent something negative (instead of to
produce something positive); or if the outcome is intangible (e.g.
recreation).

If this is the library’s first attempt at outcome-based evaluation, start
with:
Only one outcome per program
Outcomes for which you already collect indicator data
Short term outcomes
Outcomes which are easy to understand and to report to
stakeholders
 Areas in which you have experience but want to improve




Rubin for the Massachusetts Board of Library Commissioners_2004____________________________16__
So What? Using Outcome-Based Evaluation to Assess the Impact of Library Services
CONTINUUM OF PROGRAM OUTCOMES
Note that changes in individuals (outcomes) usually occur in the order
below, from bottom to top. The most difficult changes are those at
the top of the list.
Condition/ Status
Behavior
Skills
Knowledge
Community connections/ Social networks
Attitudes/ Values
Perceptions/ Feelings
Satisfaction with service
Participation/ Use of service
Awareness of service
Rhea Joyce Rubin 2003. Based on work by Patton in Utilization-Focused Evaluation (1997) and Reisman &
Clegg in Outcomes For Success! (2000).
Rubin for the Massachusetts Board of Library Commissioners_2004____________________________17__
So What? Using Outcome-Based Evaluation to Assess the Impact of Library Services
CANDIDATE OUTCOMES WORKSHEET
Users:
Community Goal:
User Need:
Library Goal:
New Service/Program:
Outputs:
Immediate or Intermediate Outcome(s):
Indicator(s):
Long-range Outcome(s):
Indicator(s):
Rubin for the Massachusetts Board of Library Commissioners_2004____________________________18__
So What? Using Outcome-Based Evaluation to Assess the Impact of Library Services
SELECTING INDICATORS AND TARGETS
FOR OUTCOME-BASED EVALUATION
SELECTING INDICATORS:
How will we know that the participants experienced the expected benefit or
change? How can we gauge the outcome or the achievement for the
participant?
An indicator is a behavior that actually demonstrates the change, or that
implies a change. An indicator identifies a statistic that summarizes the user’s
achievement of an outcome.
 One to three indicators per outcome
 Specific and unambiguous
 Measurable
 Observable and objective
 Valid
 Reliable
 Relevant
 Think of the story it tells
Rubin for the Massachusetts Board of Library Commissioners_2004____________________________19__
So What? Using Outcome-Based Evaluation to Assess the Impact of Library Services
SETTING TARGETS:
How will we know if the program has met our expectations? What is our
standard for success for the program?
 Numerical
 Realistic and based on experience (baseline figures)
 Valid
 First year/ cycle, collect data that you can use to set a target
next time. Meanwhile, use “improve” or “gain” without numerical
qualifiers
 Reasonable
AN EXAMPLE OF INDICATORS AND TARGETS:
In a job information program, an expected outcome might be that participants
learn of new job and career options.
 An indicator might be that participants can list two more
appropriate jobs/ career possibilities after the program than
before.
 A target might be that 40% of participants list at least two
more career options.
Rubin for the Massachusetts Board of Library Commissioners_2004____________________________20__
So What? Using Outcome-Based Evaluation to Assess the Impact of Library Services
DESIGNING A DATA COLLECTION PLAN
The most basic questions to consider when designing a data collection plan
are:
 What method will best capture your indicators? Since indicators
show you the change in behavior – or a behavior that implies another
type of change – you need to find the most direct way to clearly
observe and accurately count the indicators.
 What schedule will give you the most significant information?
Monthly? Annually? Note that because outcome-based evaluation is
about change, it usually requires pre- and post-tests.
Five issues to take into account:
 Collecting data should not interfere with the delivery of the service.
 Participation in the data collection activities (e.g. interview or
survey) must be voluntary and not required to receive the service.
 Participants’ privacy or confidentiality must be respected.
 The benefits of the data collection activities must outweigh the
costs for both the program and the participants.
 To minimize bias, the service provider should not serve as the
evaluator.
Four last essentials:
 Data is collected from individuals, but the results are always
aggregated and reported as a group. The idea is to measure the
outcomes of your service, not to grade an individual’s progress.
 Outcome-based evaluation of LSTA projects is not scientific
research. We have neither a scientifically selected sample of users
nor a control group. Therefore, we can make only limited claims
about our results.
Rubin for the Massachusetts Board of Library Commissioners_2004____________________________21__
So What? Using Outcome-Based Evaluation to Assess the Impact of Library Services
 Results should not be compared to results of other programs unless
the programs and the evaluation were designed and managed
together.
 The outcomes reflect the library’s contribution to a goal. But
outcomes cannot be attributed the library alone.
Rubin for the Massachusetts Board of Library Commissioners_2004____________________________22__
So What? Using Outcome-Based Evaluation to Assess the Impact of Library Services
SELECTING DATA COLLECTION METHODS
The six most common data collection methods are:






Review of existing records
Surveys
Interviews
Self-report
Observation
Tests
When determining which data collection method to use, keep several factors
in mind.
 What type of information is needed?
 How standardized must the information be?
 How valid (accurate) must the information be?
 How reliable (consistent) must the information be?
 Is this information already available somewhere? If so, who might
already have access to much, if not all, of this information?
 Is an instrument (e.g. a test or a survey form) already available? If
so, can it be modified for use in your program?
Whenever possible, use the most direct method possible. For knowledge or
skills, this may be tests or observation. For attitude change, this may be selfreport or interview. For behavior, this may be records or tests.
Review of existing records
Rubin for the Massachusetts Board of Library Commissioners_2004____________________________23__
So What? Using Outcome-Based Evaluation to Assess the Impact of Library Services
 Existing records include internal (those of your own program or agency)
and external (those kept by others) such as school or employment
records
 Internal program tracking mechanisms such as records, reports, or logs
can be used if they are consistently maintained and updated in a timely
manner and the information can be extracted easily from the system
 Do official (external) records exist that track this indicator in a way
that is useful to you? If so, are the records reliable (i.e. tracked
similarly from one individual to another)?
 Will it be possible to get the cooperation of the outside agency/
institution to get access to these official records?
 Only use external records that collect the information you really need
 Are they widely accepted as objective and credible? Be aware that
some records (e.g. program logs and attendance tallies) may not be kept
consistently and may not be directly related to achievement
Survey or questionnaire
 Good when you need information from the participant’s perspective
 Be aware that any information reported by the participant may be
inaccurate
 More systematic than anecdotal self-report (though open-ended survey
questions are also self-report)
 Can be self-administered (form filled in by participant) or asked orally
(with answers recorded by someone else)
 Easy to analyze results statistically
 Watch your language! Don’t use value-laden terms or overly personal
language
 Will participants be able to understand the survey questions? Consider
age, educational level, and cultural background.
 Will participant be able to accurately answer the questions? Consider
awareness, knowledge, and language skills.
 Are the questions culturally appropriate?
 Pilot test the survey to be sure it is easily understood and elicits the
information you need
Rubin for the Massachusetts Board of Library Commissioners_2004____________________________24__
So What? Using Outcome-Based Evaluation to Assess the Impact of Library Services
Interview
 Good when you need information from the participant’s perspective
 Must follow pre-defined questions and answers must be recorded
accurately
 Similar to oral survey but allows for follow-up questions
 Similar to self-report but more systematic
 May be used as adjunct to survey
 Labor intensive as someone must personally ask participants questions –
either by phone or in person
 Will the participant be comfortable being interviewed?
 Do you have translators as needed?
 Pilot test the interview questions to be sure it is easily understood and
that the interviewers understand when and how to ask follow-up
questions
Self-report
 Are there objective pieces of information the can be used to assess
the indicator?
 Will the participant be comfortable with self-report? What about
those who are not comfortable?
 Will the reports be acceptable as objective and accurate?
Observation
 Most appropriate when behavior change is clearly observable, i.e. a skill
or condition
 Can indicator be measured without asking questions?
 Is this indicator most accurately measured by a trained observer?
 If so, do you have staff resources to observe events and behaviors? Or
can you train volunteers?
 Beware of the participant’s privacy when using observation
Rubin for the Massachusetts Board of Library Commissioners_2004____________________________25__
So What? Using Outcome-Based Evaluation to Assess the Impact of Library Services
Cognitive or Physical Standardized Tests
 Most appropriate when the indicator is related to knowledge or skill
 Most appropriate when testing is typically included in that context, e.g.
literacy or computer skills
 Will you be able to use a pre-existing test or will you have to create
one?
 Will the participant be comfortable with testing?
 Widely accepted as credible
 Easy to analyze results statistically
 Pilot test the instrument to be sure it is easily understood and elicits
the information you need
Rubin for the Massachusetts Board of Library Commissioners_2004____________________________26__
So What? Using Outcome-Based Evaluation to Assess the Impact of Library Services
DATA COLLECTION METHODS
Review
of
Existing
Records
Written
Survey
Interview Self- Rating by
Report Trained
Observer
Low
Moderate
Moderate None
To High
Low to
Low to
Amount of Moderate Moderate
Training
Required
Moderate None
To High
Depends
Completion on
Time
Amount
of
Data
Needed
High if
Response
records
Rate
are
relevant
Long
METHOD
Cost
Moderate
To Long
Varies
Depends on
Moderate Varies
distribution* To High
Tests
Depends
Low to
on
Moderate
Cost of
Observers
Moderate Moderate
to
High
Short to
Moderate
Varies
High
High
* Surveys can be distributed on-site, by mail, or by mail with telephone follow-up. How
the survey is distributed will affect all the collection characteristics on the chart above.
For example, a survey sent by mail and returned by the participant by mail is inexpensive
but slow with a low to medium response rate depending on whether multiple mailings
and/or telephone follow-ups are done. Surveys completed on-site are fast and have a high
response rate but may be more expensive because of on-site staff costs.
Rubin for the Massachusetts Board of Library Commissioners_2004____________________________27__
So What? Using Outcome-Based Evaluation to Assess the Impact of Library Services
WRITING AN OUTCOME MEASURE STATEMENT
 Remember: outcomes are about the participants, not the program.
 Begin with only one outcome statement per program/ service.
 If you have selected a long-term outcome, you need to write at least
one intermediate outcome too.
 The participant should be the subject of the sentence. Begin the
sentence with the participant (e.g. “Preschool children…”), followed with
“will…” Who will show what?
 Outcome verbs include: know, understand, attain, increase, improve,
decrease, reduce, expand. (Note that these are not activity-based.
Activity verbs will be used in the indicators.)
 Check that the outcome is not an internal program operation (e.g.
recruiting and training staff) or an output (e.g. number of participants
served).
 Next state the indicator(s) you expect to measure plus the data
collection method you will use.
 Think SMART about your indicators: specific, measurable, attainable,
realistic, time-bound.
 Check that you have not written an objective. An objective is a SMART
statement written for traditional evaluation plans; it expresses the
projected and quantifiable products of a program. It is a statement
from the library perspective.
 Be sure it is, instead, a statement that describes what will happen to
participants as a result (at least partly) of your program. It should
express the “so what?” for your program.
Rubin for the Massachusetts Board of Library Commissioners_2004____________________________28__
So What? Using Outcome-Based Evaluation to Assess the Impact of Library Services
Rubin for the Massachusetts Board of Library Commissioners_2004____________________________29__
So What? Using Outcome-Based Evaluation to Assess the Impact of Library Services
PLANNING FOR OUTCOME-BASED
EVALUATION
 Designate an OBE coordinator to design and monitor the OBE process.
 Consider each of your library’s programs/ services. Is it aimed at
changing the lives of individual people? Is it focused more on human
impact than on internal operations? If so, it is appropriate for OBE.
 Identify the population of potential participants. It is necessary to
know for whom you are planning the service before you can consider
outcomes. Whose life will be affected by the service?
 Is there more than one intended group? Many times we can affect
people other than those we serve directly (e.g., a program designed
primarily for preschool children may affect parents as well).
 Include potential participants and others in the planning (see Selecting
Outcomes).
 Determine the program’s inputs and outputs. Consider resources and
constraints.
 Decide on program outcomes to measure. What impacts can you
reasonably expect? Are there intermediate outcomes to measure?
 Choose the indicators. Are they observable and measurable?
 Set a realistic numerical target for program success.
 Develop data collection plan (methods, timeline, and sources).
 Pretest the indicators and data collection methods. Monitor the
process re time/ staff considerations, and ease or difficulty of
Rubin for the Massachusetts Board of Library Commissioners_2004____________________________30__
So What? Using Outcome-Based Evaluation to Assess the Impact of Library Services
measurement. Tabulate and analyze results and then revise the data
collection plan as necessary.
 Write your outcome-based evaluation plan: goals, inputs, activities,
outputs, outcomes, indicators, targets, and data collection methods.
 Develop an action plan outlining what needs to be done when and by
whom.
Rubin for the Massachusetts Board of Library Commissioners_2004____________________________31__
So What? Using Outcome-Based Evaluation to Assess the Impact of Library Services
A SELECTED BIBLIOGRAPHY
Abend, Jennifer and Charles R. McClure. “Recent Views on Identifying
Impacts from Public Libraries. Public Library Quarterly 17 (3): 361-390, 1999.
Association of Specialized and Cooperative Library Agencies. “Outcomes
Evaluation Issue.” Interface 24(4):1-11, Winter 2002.
Bertot, John Carlo and Charles R. McClure. “Outcomes Assessment in the
Networked Environment: Research Questions, Issues, Considerations, and
Moving Forward.” Library Trends 51 (4): 590-613, Spring 2003.
Corporation for National and Community Service. AmeriCorps Program
Applicant Performance Measure Toolkit. February 14, 2003. On the web at
http://www.projectstar.org
Dervin, Brenda and Benson Fraser. How Libraries Help. Stockton, CA:
University of the Pacific, 1985.
Durrance, Joan and Karen E. Fisher. “Determining How Libraries and
Librarians Help.” Library Trends 51 (4): 541-570, Spring 2003.
Durrance, Joan C. and Karen E. Fisher-Pettigrew. How Libraries and Librarians
Help: Outcome Evaluation Toolkit. October 2002. On the web at
http://www.siumich.edu/libhelp/toolkit
Durrance, Joan C. and Karen Pettigrew. How Libraries and Librarians Help.
Chicago: ALA Editions, forthcoming 2004.
Durrance, Joan C. and Karen E. Fisher-Pettigrew. “Towards Developing
Measures of the Impact of Library and Information Services.” Reference and
User Services Quarterly 42(1): 43-53, Fall 2002.
Hernon, Peter and Robert E. Dugan. An Action Plan for Outcomes Assessment
in Your Library. Chicago: ALA, 2002.
Rubin for the Massachusetts Board of Library Commissioners_2004____________________________32__
So What? Using Outcome-Based Evaluation to Assess the Impact of Library Services
Hernon, Peter and Robert E. Dugan. Outcomes Assessment: Views and
Perspectives. Westport, CT: Libraries Unlimited, forthcoming 2004.
Hernon, Peter. “Outcomes Are Key But Not the Whole Story.” Journal of
Academic Librarianship. 28(1-2): 54-55, 2002.
Horn, Claudia B. Outcome-Based Evaluation for Literacy Programs. Syracuse,
NY: Literacy Volunteers of America, Inc., 2001.
Innovation Network, Inc. Helping NonProfit Agencies Succeed Their
workstation is a free web-based interactive planning tool that assists you to
write an outcomes-based evaluation plan while you are online. They estimate 2
to 4 hours to write a plan but it can be done in short steps; all your
information is stored for your future use. 1992. On the web at
http://www.innonet.org/workstation/workstation_handbook.pdf
Institute of Museum and Library Services. Frequently Asked OBE Questions.
Washington, DC: IMLS, rev. 2002. In print and on the web at
http://www.imls.gov/grants/current/crnt_outcomes.htm
Institute of Museum and Library Services. New Directives, New Directions:
Documenting Outcomes. Washington, DC: IMLS, rev. 2002. In print and on the
web at http://www.imls.gov/grants/current/crnt_obebasics.htm
Institute of Museum and Library Services. Perspectives on Outcome Based
Evaluation for Libraries and Museums. Washington, DC: IMLS, 2000.
Kellogg, W.K. Foundation. Evaluation Handbook. (See especially pages 28-46 on
OM. ) Battle Creek, MI. 1998. Also available on the web at
www.wkkf.org/pubs/Pub770.pdf
Kellogg, W.K. Foundation. Guiding Program Direction with Logic Models. Battle
Creek, MI. 2003. In print and available on the web at
www.wkkf.org/pubs/tools/evaluation/LMDGsummary_00447_03674.pdf
Rubin for the Massachusetts Board of Library Commissioners_2004____________________________33__
So What? Using Outcome-Based Evaluation to Assess the Impact of Library Services
Kellogg, W.K. Foundation. Logic Model Development Guide: Using Logic Models
to Bring Together Planning, Evaluation, and Action. Battle Creek, MI. 2001.
Also available on the web at www.wkkf.org/pubs/Pub1209.pdf
Lance, Keith Curry and others. Counting on Results: New Tools for OutcomeBased Evaluation of Public Libraries. (Final report). Washington, DC: IMLS,
2002. On the web at www.lrs.org
Marshall, Joanne Gard. “Determining Our Worth, Communicating Our Value.”
Library Journal 125 (19):28-30, November 15, 2000.
Matthews, Joseph R. Measuring for Results: The Dimensions of Public Library
Effectiveness. Westport, CT: Libraries Unlimited, 2004.
McNamara, Carter. “Basic Guide to Outcomes-Based Evaluation for Nonprofit
Organizations with Very Limited Resources.” For the Management Assistance
Program for Nonprofits. On the web at
www.mapnp.org/library/evaluatn/outcomes.htm
Mika, Kristine L. Program Outcome Evaluation: A Step-by-Step Handbook.
Milwaukee: Families International, 1996.
Morley, Elaine and others. Outcome Measurement in Nonprofit
Organizations:Current Practices and Recommendations. Washington, DC: The
Independent Sector, 2001.
Organizational Research Services. Outcomes for Success 2000! Seattle, WA:
The Evaluation Forum, 2000.
Organizational Research Services. Managing the Transition to OutcomesBased Planning and Evaluation. Seattle, WA: The Evaluation Forum, 1998.
Patton, Michael Quinn. Qualitative Research and Evaluation Methods. 3rd
edition. See pages 151-159 and 518-534. Thousand Oaks, CA: Sage
Publications, 2002.
Rubin for the Massachusetts Board of Library Commissioners_2004____________________________34__
So What? Using Outcome-Based Evaluation to Assess the Impact of Library Services
Patton, Michael Quinn. Utilization-Focused Evaluation. 3rd edition. See
especially pages 147-177. Thousand Oaks, CA: Sage Publications, 1997.
Poll, Roswitha. “Measuring Impact and Outcome of Libraries.” Performance
Measurement and Metrics 4 (1): 5-12, 2003.
Project Literacy Victoria. Outcome Measurement for a Community Literacy
Program. Victoria: British Columbia, 2001 and on the web at:
http://www.plv.bc.ca/outcomes/programoutcomes.pdf
Project STAR for the Corporation for National Service. Programming for
Impact Manual. On the web at http://www.projectstar.org
Reisman, Jane. A Field Guide to Outcome-Based Program Evaluation. Seattle,
WA: The Evaluation Forum, 1994.
Rubin, Rhea Joyce. “Predicting Outcomes: Outcome Measurement as a Planning
Tool” in Joan C. Durrance and Karen Pettigrew. How Libraries and Librarians
Help. Chicago: ALA Editions, forthcoming 2004.
Sadlon & Associates, Inc. Workbook: Outcome Measurement of Library
Programs. Tallahassee: State Library of Florida, 2000. On the web at
http://dlis.dos.state.fl.us/bld/research_office/bld/research.htm
Schalock, Robert L. Outcome-Based Evaluation. NY: Plenum Press, 1995.
Steffen, Nicolle O., Keith Curry Lance and Rochelle Logan. “Outcome-Based
Evaluation and Demographics for the Counting on Results Project.” Public
Libraries, 41(4): 222-228. July-August 2002.
Steffen, Nicolle O., Keith Curry Lance and Rochelle Logan. “Time to Tell the
Whole Story: Outcome-Based Evaluation and the Counting on Results Project.”
Public Libraries, 41(5): 271-280. September-October 2002.
Rubin for the Massachusetts Board of Library Commissioners_2004____________________________35__
So What? Using Outcome-Based Evaluation to Assess the Impact of Library Services
Taylor-Powell, Ellen and others for the University of Wisconsin Extension.
Planning a Program Evaluation. A series of excellent short pamphlets, 19962002. In print and on the web at http://www.uwex.edu/ces/pdande/evaluation
United Way of America. Agency Experiences with Outcome Measurement:
Survey Findings. Alexandria, VA: United Way, 2000.
United Way of America. Measuring Program Outcomes: A Practical Approach.
Alexandria, VA: United Way, 1996.
United Way of America Outcome Measurement Network. On the web at
http://national.unitedway.org/outcomes
Williams, Harold S. and others. Outcome Funding: A New Approach to
Targeted Grantmaking. 3d. ed. Rensselaerville, NY: The Rensselaerville
Institute, 1995.
Rubin for the Massachusetts Board of Library Commissioners_2004____________________________36__
So What? Using Outcome-Based Evaluation to Assess the Impact of Library Services
DATA COLLECTION AND ANALYSIS:
A SELECTIVE & DESCRIPTIVE BIBLIOGRAPHY
Bolan, Marc and others. How to Manage and Analyze Data for Outcome-Based
Evaluation. Seattle, WA: The Evaluation Forum, 2000. How to manage
program data with specifics on how to use Access and Excel for data
preparation and analysis. This clearly written book, which comes with a
diskette of fictitious program data for practice purposes, may be more than
you need to know if you are running a small program with few participants.
Bond, Sally, Sally Bond, and Kathleen Rapp. Taking Stock: A Practical Guide to
Evaluating Your Own Programs. Chapel Hill, NC: Horizon Research Inc., 1997.
Although this slim volume (93 pages) covers more than data collection, the
sections on that topic are well-presented for novices.
Fink, Arlene and Jacqueline Kosecoff. How to Conduct Surveys: A Step-byStep Guide. 2d. ed. Thousand Oaks, CA: Sage Publications, 1998. A practical
and concise (111 pages) paperback guide.
Fink, Arlene and others. The Survey Kit. 2d ed. Thousand Oaks, CA: Sage
Publications, 2002. If Fink’s guide, above, is not enough for you, this newlyupdated ten volume set should answer all your questions about written surveys
and interviews. Each volume (100 to 200 pages) focuses on a specific subject:
self-administered & mail surveys, in-person interviews, telephone interviews,
sampling, reliability & validity, data analysis, and reports. The first two
volumes, “The Survey Handbook” and “How to Ask Survey Questions” are
especially useful to new survey developers.
Glitz, Beryl. Focus Groups for Libraries. Chicago: Medical Library Association,
1999.
Henerson, Marelene E. and others. How to Measure Attitudes. Newsbury Park,
CA: Sage Publications, 1987. Attitudes, unlike skills or knowledge, cannot be
Rubin for the Massachusetts Board of Library Commissioners_2004____________________________37__
So What? Using Outcome-Based Evaluation to Assess the Impact of Library Services
observed directly; they are therefore more difficult to measure. This is an
excellent guide.
Herman, Joan and others. Program Evaluation Kit. Thousand Oaks, CA: Sage
Publications, 2d. ed. 1987. A nine volume set covering all facets of program
evaluation. Each paperback volume is medium length (96 to 192 pages) on one
of the following topics: how to focus an evaluation, how to design, how to use
qualitative methods, how to assess program implementation, how to measure
attitudes, how to measure performance tests, how to analyze data, how to
communicate evaluation findings. Each volume can stand alone and may be
purchased separately.
Insititute of Museum and Library Services. Introduction to Outcome
Oriented Evaluation: Selected Resources. Washington, DC: IMLS, February,
2002. On the web at www.imls.gov/grants/current/crnt_bib.html Lists
materials on user surveys and other data collection tools and methods used by
museums, science programs, and other non-profit and educational agencies.
Also lists automated online survey tools.
Krueger, Richard A. and David L. Morgan, eds. The Focus Group Kit. Thousand
Oaks, CA: Sage Publications, 1998. A six volume set covering everything re
focus groups. Each volume is a short (100 pages), clear paperback on one of
the following focus group topics: planning, developing questions, involving
community members, moderating, analyzing & reporting results. Each volume
can stand alone and may be purchased separately.
Peterson, Robert A. Constructing Effective Questionnaires. Thousand Oaks,
CA: Sage Publications, 1999. Another short (152 pages) manual for the new
evaluator.
Resiman, Jane and others. A Field Guide to Outcome-Based Program
Evaluation. Seattle, WA: The Evaluation Forum, 1994. After a 30 page
introduction to OM concepts, the next 75 pages of this book introduce data
methods with specifics on formats, sampling, and statistical analysis for
novices.
Rubin for the Massachusetts Board of Library Commissioners_2004____________________________38__
So What? Using Outcome-Based Evaluation to Assess the Impact of Library Services
Reisman, Jane. How to Manage and Analyze Data for Outcomes-Based
Evaluation. Seattle, WA: The Evaluation Forum, 2000.
Salkind, Neil J. Statistics for People Who Think They) Hate Statistics.
Thousand Oaks, CA: Sage Publications, 2000. As the title suggests, an
introduction to statistics for novices.
Simon, Judith Sharken. Conducting Successful Focus Groups: How to Get the
Information You Need to Make Smart Decisions. St. Paul, MN: Amherst H.
Wilder Foundation, 1999. A clear, concise, short (67 page) introduction to
how and when to use focus groups.
Zemke, Ron and Thomas Kramlinger. Figuring Things Out. Reading, MA:
Addison-WesleyPublishers, 1982. Written for trainers, this book has strong
introductions to focus groups, interviews, gathering information, surveys and
questionnaires, and sampling. I recommend those sections despite its 1982
publication date.
Zweizig, Douglas and others. The Tell It! Manual: The Complete Program for
Evaluating Library Performance. Chicago: ALA, 1996. Another book on a
broader topic that includes excellent chapters on data collection methods.
See Part III, which has 9 chapters (111 pages) on surveying, interviewing,
attitude measurement and more.
Rubin for the Massachusetts Board of Library Commissioners_2004____________________________39__
Download