Project Snowball – sharing data for cross-institutional benchmarking Lisa Colledge, Anna Clements, M’hamed el Aisati, Scott Rutherford euroCRIS 2012 with modifications for JISC RIM Meeting Bristol th Jun 28 2012 The aims of Snowball • Higher education institutions – Agree a common set of metrics to support institutional decision making – Reach consensus on standard methodologies for calculating these metrics – Publish the “recipe book” as open standard definitions • These metrics will cover the entire landscape of research activity • These metrics will become global sector standards BACKGROUND The origins of these aims... • Growing recognition of value of metrics to support strategies • Dissatisfaction with the tools available • Frustration over availability of metrics to make sensible measurements Joint Imperial-Elsevier JISC-funded study of research information management, available via http://www.projectsnowball.info/ RECOMMENDATIONS • Institutions and funders should work more collaboratively, and develop stronger relationships with suppliers • An agreed national framework for data and metric standards is needed • Suppliers should participate in the development of data and metric standards Snowball has evolved from these recommendations The goal? To enable crossinstitutional benchmarking • Agree methodologies for a standard set of metrics to support strategic decision making • Driven by higher education institutions with recognised common challenges and goal - working with a supplier (Elsevier), with everyone contributing voluntarily Comprehensive metrics landscape Metrics require institutional, proprietary and third party data Test 1 to calculate the metrics landscape Approach: institution and Elsevier contribute data on 10 chemistry researchers as proxy for the whole university Definitions of metrics Data availability across landscape Sensitivity of some data types (next slide) Researcher-level data Manual labour in data collection Data types with high sensitivity Test 2 of metric calculation feasibility Approach: institution and Elsevier test scalability by contributing data on whole university for a smaller set of metrics Definitions of metrics Data availability across landscape Experts group formed to select and define phase 1 metrics – impactful, doable, require data from 3 sources Sensitivity of some data types Data agreement prepared by partners Most sensitive data types not phase 1 Researcher-level data Used minimally Metric granularity Manual labour in data collection Institution and Elsevier supply data as close to native as possible Test 2 of metric calculation feasibility Test 2 of metric calculation feasibility HESA FTE research, reserch & teaching HESA cost centre Metrics require institutional, proprietary and third party data Project Snowball recap • Driven by sector • Facilitated and supported by Elsevier • Public service The project has demonstrated feasibility of scalably inputting data from 3 sources to generate metrics and benchmarks • Institutional • Proprietary • Third party Next steps • Publish the phase 1 metrics “recipe book” as open standards – Sep 2012 • Refine phase 1 metrics as global standards, and extend same approach to more metrics • CERIFy metrics – meeting scheduled Sep 2012 • Spread the word – Russell Group, 94 Group, Vendors, Funders Thank you for your attention www.projectsnowball.info l.colledge@elsevier.com j.green@imperial.ac.uk anna.clements@st-andrews.ac.uk