DEVELOPING BEST-FIT EVALUATION STRATEGIES

advertisement
Best-Fit Evaluation Strategies:
Are They Possible?
John Carlo Bertot, John T. Snead,
& Charles R. McClure
Information Use Management and Policy Institute
College of Information, Florida State University
http://www.ii.fsu.edu
Introduction

Received a grant from the Institute of Museum and Library
Services (IMLS) to develop an online instructional system to
assist public libraries evaluate their services/resources



Evaluation Decision Management System (EDMS)
In year two of three year project
Why this project?




Little research provided comprehensive assistance in determining
what specific evaluation strategies serve libraries best relative to
specific library situational factors and contexts, data needs, and
other considerations
With so many evaluation options available, there was a need to
bridge information-need issues (i.e., situational factors, data needs,
stakeholder questions, etc.) with evaluation approaches
Understanding information needs and linking these needs to
evaluation approaches required evaluation strategies capable of
providing data that library decision makers can use to address
specific problems
Bridge practice an research
Best Fit Evaluation Approaches

The original intent of the project was to
develop


Best-fit evaluation strategies which match the data
needs of a library within a specific situational
context to the evaluation approaches that are
most appropriate to that particular situational
context
Identify which evaluation strategies supply the
best data within specific library organizational and
situational contexts for use to provide the greatest
impact to improve library services, or enable
libraries to better advocate the value of libraries to
their institutions or the communities they serve
Best-Fit Evaluation Strategy
Considerations





What evaluation approaches are available
Which evaluation approaches might best meet a
library’s data needs, either library developed or
imposed by external funders/organizations/etc.
How to develop an overall evaluation plan that makes
effective and efficient use of limited library resources
How to implement an evaluation strategy
How to use evaluation findings to advocate for local
library support
The Original Approach

Academic Literature








Outcomes assessment
Value demonstration
ROI
Outputs
Service Quality
Other
Stand alone literature and approaches
Bring these to the public library community
Different Perspectives

Stakeholder Perspective



Answer a range of questions asked by various stakeholders
groups (e.g., library boards, county or city executives)
regarding library services and resources
Make informed decisions regarding a library’s range or
availability of services and resources
Data Perspective, provide data to





Answer a range of questions asked by various stakeholders groups (usercentered evaluation perspective)
Make informed decisions regarding a library’s range or availability of services
and resources (library-centered evaluation perspective)
Demonstrate value and effectiveness of the library to the community that it
serves (community-centered evaluation perspective)
Frame the perceptions of the library in the local political environment (political
context-centered evaluation perspective)
Support the notion of the library as serving as a public good (customercentered evaluation perspective)
Then the Reality of Public
Libraries


Advocacy
Tell story





Value
Impact
Cuts across individual evaluation
approaches
Integrated
Fits library’s situation
Development of Best-Fit
Evaluation Strategies



Library decision makers are often faced with
difficulties matching their data needs with the
appropriate evaluation approaches
There are many different kinds of evaluation
data that a library may need and evaluation
approaches that a library might employ
As a result, many libraries struggle with the
problem of choosing the best evaluation
approaches to effectively and efficiently
demonstrate the value they provide

And…they don’t necessarily consider “evaluation”
Some Issues and Findings

Have not quite abandoned the “logic model”
approach to evaluation and “best fit”, but
have


Modified thinking significantly
Changed how we make evaluation tools available


Less “evaluation like”
Focused more on advocacy and story telling -with evaluation on the side

Though it does underlie the entire approach
Some Issues and Findings

In essence, developing a community of
practice



Resources
Access to experts
Information commons for asking questions and
sharing ideas, resources, and other material



Successes and failures
How to
Other
Some Issues and Findings

Public libraries don’t necessarily consider evaluation as part of their
advocacy strategy (if they have one)


The link between evaluation, data, and advocacy is not always apparent
Once burned, twice shy

If have used evaluation strategies and data collection efforts, and not met
with success, there is a reluctance to do so again



Library situational factors (organizational, community, other) affect the
successful use (or unsuccessful use) of leading evaluation approaches
Though libraries now talk about advocacy regularly -- the extent to
which they are able and/or ready to engage in advocacy strategies is in
question


Regardless of a linkage between evaluation and advocacy
The fear is that we may have to hide evaluation in an advocacy
wrapper


Waste of resources
These are two different things and require different skill sets
The ability of a library director to negotiate the political
environment of the library may trump evaluation
Some Issues and Findings

Need to provide as many tools as possible to
facilitate evaluation

If, then interfaces that enable EDMS to know
something about libraries



NCES data (operating budgets, population served, FTEs,
other)
Public Library Internet data (connectivity, bandwidth,
workstations, services, other)
Prepackaged reports which incorporate data



Peer comparisons
Brochures
Presentations (Budget, building, other)
Moving Forward

Challenges remain



Biggest is making link between evaluation
and advocacy
“Painless” evaluation that meets library’s
situational factors
Redefining success of project
Conclusions




Some libraries do continue to engage in a wide range of
evaluation efforts to assess services and resources provided to
the communities libraries serve
The evaluation environment is increasingly complex, and
requires knowledge of multiple evaluation frameworks,
methodologies, data analysis techniques, and communication
skills
The issue is not that libraries face a lack of available evaluation
approaches
The issue is helping libraries select the approach or approaches
that best meet advocacy needs of a library from the many
evaluation techniques that exist

Making the evaluation-advocacy link
Thank You

Questions/Comments?

Contact information

John Bertot: jbertot@fsu.edu

Information Institute: http://www.ii.fsu.edu
Download