versus Count-Ability Denise Troll Covey Associate University Librarian, Carnegie Mellon NERCOMP Library/IT SIG Workshop – April 2002 Personal Observation • Lots of data gathering • Some data are compiled • Some compiled data are analyzed • Some analyzed data are actually used Wasted effort? Personal Suspicion • Lots of data languishing – Why? – Accountability for data reporting is minimal – Data management is cumbersome – Sense that some of the data aren’t useful – Imperative to implement new measures New Data Gathering • Evidence of becoming “Predominantly digital” • Address Advisory Board concerns 100% Traditional 80% Digital 60% 40% 20% 0% Visits Reserves Reference Assessment Challenges • Deciding what data to gather • Understanding the whole assessment process • Determining what skills to acquire & how • Organizing assessment as a core activity • Managing assessment data • Acquiring an interpretive context Distinguished Fellowship Nov 2000 – Oct 2001 2001 MIS Task Force • Assess current & proposed data practices • Determine what data to gather & how • Develop a specification for a new MIS that solves current problems • Oversee implementation & deployment of the new MIS MIS TF Time Line INITIAL REVISED Conduct data audit & needs assessment Recommend data to gather & manage in MIS May 01 Jul 01 Jun 01 Apr 02 Prepare & approve functional specification Evaluate & recommend software for new MIS Prepare & approve design specification July 01 Sept 01 Nov 01 Jun 02 Aug 02 Nov 02 Implement & test MIS prototype Implement & test production MIS Feb 02 May 02 Feb 03 May 03 Document MIS & provide training Migrate data & release new MIS Jun 02 Jul 02 Jul 03 Aug 03 May-June 2001 – Data Audit • Interviews with functional groups – What data do you gather or plan to gather? – How do you gather the data? How often? – Who uses the data? How often? – For what purpose do you gather the data? – For what audience do you gather the data? – How do you present the data? To whom? – How do you store, access, & manipulate the data? – What problems do you have with data? July 2001 – Assemble Results • Created 12-page spreadsheet – data, department, source, audience, software, process, problems, & solutions • Identified common problems – Data management is decentralized & uncoordinated – Current “MIS” is incomplete & not kept up-to-date – Errors or anomalies in the data are not corrected – Data gathering & entry are too complicated – Difficult to generate multi-year trend lines – Sheer volume of data is overwhelming More Problems • Lack of communication, training, & support – What data are available? Where? – How do I access & use the current MIS? – Problems with access privileges, server, & network • Wasted resources – Duplicate efforts – No one has time, skills, or responsibility to analyze the data – Data are either not used or under used August 2001 – Libraries Council • Distributed spreadsheet to governing council – To confirm information was accurate & complete – To confirm problems with current data practices – Are we gathering the right data? – Is the data we’re gathering worth the effort? • Outcome – Add missing data (e.g., ACRL, IPEDS) – Add more proposed new measures Sept 2001 – Field Trip • Keep task force small & strategic – Members with power, willing to take the heat & do the work • Be judgmental about usefulness of the data • Distinguish data for external & internal audiences • Data gathering advice – Stop counting what you don’t need – Automate everything you can – Use more sampling • Identify data for MIS based on research questions October 2001 – Department Heads • Added data LC requested • Distributed 15-page spreadsheet organized by department – To confirm information was accurate & complete • Outcome: clarification of minor details November 2001 – Winnow the List • Least useful data = data required for ACRL & IPEDS – University Librarian will consider NOT gathering the data we don’t really use it or it can’t be automated • Don’t know enough about data gathering & use – Someone above them in the food chain uses the data – Data definitions are inconsistent or shifting – How difficult is it to gather data manually? – Can data gathering be automated? IF February 2002 – Department Heads • Distributed spreadsheet per department – data (current & proposed), audience (internal & external) – How used by department? – Is data gathering automated? Can it be automated? – Recommend – Keep regularly, Sample, or Not gather • Outcome – Few recommendations to Sample or Not gather data – Uncertainty about what was or could be automated – Hypotheses about how data could be used Current Data Gathering Administration Access Services Acquisitions Cataloging Science Libraries Hunt Reference Arts & Spec Coll Archives & DLI LIT Operations LIT R&D TOTAL Sample 1 2 1 1 3 3 6 1 1 5 7 Keep routinely 0 17 Sample Increase 3 3 1 6 1 1 3 3 6 0 1 5 27 % Increase Sample Keep routinely Manual % Increase Manual Automated % Increase Proposed Data Gathering GRAND TOTAL Proposed Data Gathering 12 32 16 12 8 9 17 7 7 9 129 8% 19% 6% 8% 38% 33% 35% 100% 83% 100% 100% 100% 100% 100% 0% 50% 0% 0% 0% 0% 0% 14% 56% 21% 0% 0% 74% 0% 0% 11% April 2001 – Recommendations to LC • What data to gather • How to gather the data • What data to manage in 1st version of new MIS Criteria for Data Gathering • Benefits must be worth the cost • Gather data that are useful & easy to gather – Usefulness gauged by relationship to strategic & digital plans, & Advisory Board concerns – Keep regularly data that are or can be gathered automatically – Sample data gathered manually, used less frequently, only by internal audiences • Don’t gather data that are not useful or are difficult to gather & What Data to Gather & How C = Current P = Proposed STRATEGY Eliminate C 11 P 6 TOTAL 17 15% Sample Keep regularly Already automated Automate if possible Manual TOTAL to gather 13 3 16 22 13 36 84 14% 22 19% 7 20 17% 6 42 36% 16 100 Changes in Data Gathering 125 Eliminate Sample Automated Manual 100 75 50 25 0 Current Proposed Recommended Criteria for Data in the New MIS • Data gathered regularly & useful with external audiences – University administrators, ACRL, IPEDS, & FactBook – Measures of important traditional or digital trends – Selected service costs & expenditures • Small enough data set to implement in a year – Other data may be added to MIS later – All data gathered will NOT be managed by the MIS • For example, department goals for Advisory Board Next Steps for MIS TF May 2002 Decide what data manipulations, access methods & controls, & graphics we want the new MIS to do Consider additional new measures July 2002 Determine the feasibility of what we want Document functional requirements specification Submit to the Libraries Council (LC) for approval Sept 2002 Evaluate software & make recommendation to LC Dec 2002 Design the user interface of the new MIS Use paper prototyping to facilitate design work Jan 2003 Begin implementing new MIS prototype Next Steps for MIS TF ??? 2-3 weeks MIS prototype ready for data entry & testing Test prototype MIS ??? Revise design & functionality based on testing ??? Implement MIS Work with LC to decide what existing data, if any, gets “migrated” to the new MIS 2-3 months Documentation, training, data entry, & testing ??? New MIS released “Culture of Assessment” • Traditional & emerging measures – Inputs, outputs, outcomes, composites, performance – Meta-assessments of new measures • Guidelines, best practices, & standards • BS about “creating a culture” – As if new know-what & know-how are enough – No attention to what a culture really is – No understanding of what it takes to change a culture What is a Culture? • Beliefs – values & expectations • Behaviors – observable activities • Assumptions – unconscious rationale for continuing beliefs & behaviors Conner, D.R. Managing at the Speed of Change. NY: Villard, 1992. Orchestrating a Culture • Conduct a cultural audit – If there’s a gap between your current culture & your objectives, close it – If you don’t close it, the current culture wins CURRENT Beliefs Behaviors Assumptions Transition DESIRED Beliefs Behaviors Assumptions Audit via Observation CURRENT DESIRED Beliefs • Data aren’t useful • Data are useful • Data aren’t used • Data are used • No reward for work • Work is rewarded Behaviors • Haphazard data gathering, • Accurate, timely data reporting, compilation, & gathering, reporting, analysis compilation, & analysis • Inefficient data mgmt • Efficient data mgmt • Ineffective data use • Effective data use Assumptions Data aren’t important Data are important Audit via Questionnaire • Survey of perceptions of assessment practices, priorities, & problems – Perception by level in organizational hierarchy • Administrator, Middle manager, Other – Perception by library unit • Public services, Technical services, IT – Perception by status • Faculty, Staff Audit via Questionnaire • Somewhat agree on top three assessment priorities – Understand user behaviors, needs, expectations, & priorities – Operate the libraries cost-effectively – Validate expenditures • Disagree on – How assessment is organized – What assessments are conducted – What resources are allocated to assessment – What we need to solve our assessment problems Creating a Culture of Assessment • Change beliefs, behaviors, & assumptions – Resistance is natural, but it should be futile – The onus is on administration & management • To apply consequences (motivate) • To provide training (enable) • Change management is pain management – Pain = incentive to disengage from the status quo – Remedies = incentive to adopt the vision & plans Communication is Critical • Administrators must focus & secure commitment – Articulate goals & rationale – Develop & distribute action plans – Manage expectations – Assign responsibilities – Provide training & consequences – Answer questions – Identify & solve problems – Repeat as necessary The alternative is to suffer the consequences of resistance, failure, or missed opportunity Speed of Technological Change 70 60 50 40 30 20 10 0 Change is inevitable & accelerating By 2015, it will occur at an exponential rate 1960 1970 1980 1990 2000 2010 2020 Good management without sound administration is like straightening the deck chairs on the Titanic Thank You Denise Troll Covey Associate University Librarian Carnegie Mellon troll@andrew.cmu.edu 412-268-8599