RBMS Metrics TF Group 4 – Web Usage – Mark draft v2 130121

advertisement
RBMS Metrics and Assessment Task Force
Subgroup 4: Use & Users, Web Stats, Circulation, Gate Counts, and Community Impact
Report Section Draft – Mark Greenberg (01/21/13)
WEBSITE USAGE
Web analytics data for special collections websites, including “home pages” describing
department and collections/services overall, virtual exhibits, collection databases (finding aids,
image databases, etc.); statistics from social media applications; user surveys).
Measuring Operational Activities, Efficiencies, etc.:
1. Describe the activity or process to be measured.
The activity measured consists of patron access to web-based special and digital collections
content, including department and collection information, services, digital exhibitions, policies
and procedures, and blogs on departmental websites, in institutional repositories, in commercial
repositories, and/or on social media sites. Digital objects created, converted, or born-digital may
include text (including ETDs), images, audio/video, moving image materials, and EADs.
2. Identify existing metrics (i.e., measures or counts), noting any references to
international, national, or community standards, guidelines, or definitions.
Existing metrics for web page use offer a wide range of information, including patron
demographics, the total number of distinct web pages and digital objects publicly available to
patrons on a special and/or digital collections website; the total number of times patrons view
and/or provide feedback for each page; the time spent on each page; and the number of times
patrons view, provide feedback, and/or download digital objects. Facebook has its own metrics
and definitions, including total likes, friends or fans, posts, people talking about this, check-ins,
and weekly total reach.1 YouTube metrics bear a close resemblance to Google Analytics, with
appropriate modifications for video content.2
1
2
See: http://ads.ak.facebook.com/ads/creative/insights/page-insights-guide.pdf
See: http://support.google.com/youtube/bin/answer.py?hl=en&answer=1714323
RBMS Metrics Task Force, Subgroup 4
Website Usage / 2 of 5
Prior to recent changes, ARL Supplementary Statistics counted the following relevant statistics:
number of search queries in databases or services, number of virtual visits to the library website
and catalog, number (objects or unique files) and size (in gigabytes) of digital collections,
number of times digital items were accessed, and number of queries conducted.3 ARL defined
virtual visits to the library website and catalog as “a user’s request of the library website or
catalog from outside the library building excluding the number of pages or gratuitous elements
… viewed” and excluded robot or spider crawls and page reloads. For digital objects and files,
ARL also excluded backup copies or mirror sites in order to prevent counting the same digital
content more than once. The 2011-12 ARL Statistical Survey removes data for virtual website
visits but in the section “Titles and Volumes,” it states “If your library digitizes content from its
own collection and the content is accessible under current copyright law you can report it. Do not
count HathiTrust, CRL, Internet Archive, etc. unless your library owns the digitized item and it
is accessible under current copyright law.”4 Further, for 2011-12 statistics, ARL has asked
library members to report separately for Special Collections units, total expenditures (excluding
fringe benefits), total materials expenditures, other operating expenditures, and total staff FTE.
ARL offers extensive instructions for reporting expenditures salaries and wages, and staffing.
ACRL also collects statistics. Its 2010-2011 worksheet requests information on the number of
virtual visits to the library website and catalog and uses the ARL data collection instructions.
ACRL also asks for the number of digital collections and items, their total size in gigabytes, the
number of times items were accessed, the total number of queries conducted, and the direct cost
in personnel, equipment, software, and contracted services to construct and manage digital
collections.5
Additional information on existing metrics is provided by a variety of digital content
repositories. The Florida Virtual Campus counts a view every time a user opens an object. A
search query that produces results does not count as a view. Accessing the bibliographic record
for an object also does not count as a view. Anytime the user views an item in the Image Viewer,
PDF viewer, etc., it counts as one view. For digital books, serials, etc., that display in the Image
Viewer, each time the user “turns the page,” it counts as a view. Multi-page texts packaged as a
PDF count as one view no matter the number of pages turned.
Publicly accessible website metrics or statistics for special and digital collections sites
specifically proved difficult to find and may not be reported separately at most institutions, but
3
See, for example, the 2008-09 ARL Supplementary Statistics Worksheet, available from:
http://www.libqual.org/documents/admin/09ssurvey.pdf
4
See the instructions for completing the 2011-12 ARL Statistical Survey, available from:
http://www.arlstatistics.org/About/Mailings/stats_2011-12
5
See: http://acrl.countingopinions.com/docs/acrl/ACRL11Survey.pdf
RBMS Metrics Task Force, Subgroup 4
Website Usage / 3 of 5
library-wide statistics are available on the Internet for several institutions. Metrics included in
their online reports vary. For example, Georgetown University Library for FY 2012 counted
unique visitors, web pages viewed, home page viewed, and top visited web pages.6 The
University of Georgia Libraries reports total page views and total visitors.7 Other institutions,
such as the Augustus C. Long Health Sciences Library at Columbia University, reported total
page views, visitors, access paths, and browser and mobile device use.8 Duke University
Libraries reports visits and page views, campus vs. off campus access, how patrons found the
libraries’ websites, and browser and mobile device use.9
3. Comment on:
a. the adequacy and the inadequacy of existing metrics, especially those that reference
international, national, or community standards, guidelines, or definitions;
A modest body of scholarship within the last decade analyzes best practices in library website
metrics and analysis.10 Several institutional websites offer additional insight. Iowa Library
Services, State Library of Iowa that the number of visits and pages visited are the most valuable
web metrics, and recommends “pay little or no attention to the other statistics.” Iowa Library
Services defines a visit as “every time a browser open your Website.” A unique visitor is counted
as one computer or IP address visiting the website,11 but ARL noted the misleading nature of
some user’s IP address due to ISPs, firewalls, and proxy servers. According to Thomas Crichlow,
Duke University Libraries’ varied collections, multiple website domains, and diverse patron
usage patterns across domains, make accurate, broad generations difficult. Changes to Duke’s
digital asset management system saw total page views fall, but 87 percent of the decrease was
due to changes to how search results display, multi-item view sizes, and paginated items.12
Less clear from the metrics is how best to interpret patron behavior in the context of the content
accessed. For example, does “average time on page” have the same meaning for a home page
and digital exhibition page? Time spent carefully reading content within a digital exhibition
likely is far more meaningful to research and learning than patrons who struggle to find
appropriate links on a home page and thus linger. Similar issues exist for a variety of other
metrics, and thus content context is important to assessment.
6
See: http://www.library.georgetown.edu/about-us/numbers/web
See: http://www.libs.uga.edu/statistics/
8
See: http://library.cumc.columbia.edu/health-sciences-library-website-statistics
9
See: http://library.duke.edu/about/assessment/web/documents/webstats-fall-2011.pdf
10
See in particular, Elizabeth L. Black, “Web Analytics: A Picture of the Academic Library User,” Journal of Web
Librarianship, vol. 3, no. 1, pp. 3-14; Laura B. Cohen, “A Two-Tiered Model for Analyzing Library Website Usage
Statistics, Part 1: Web Server Logs,” Portal: Libraries and the Academy, vol. 3, no. 2, pp. 315-26, and “Part 2: Log
File Analysis,” Portal: Libraries and the Academy, vol. 3, no. 3, pp. 517-26.
11
See: http://www.statelibraryofiowa.org/ld/q-s/silo/web/web-stats/web-stats-intro
12
See: http://library.duke.edu/about/assessment/web/documents/webstats-fall-2011.pdf
7
RBMS Metrics Task Force, Subgroup 4
Website Usage / 4 of 5
b. methods commonly used to collect data required by the metric;
Google Analytics (GA) offers an effective page tagging tool for keeping statistics on website
usage, including for tools like LibGuides, though server log file products such as AWStats and
Webalizer are also useful. GA collects a wide range of information, including information on
user demographics, behavior, technology, mobile devices, visitor flow, and content. Metrics
include page views, unique page views, pages viewed per visit, average time on page, bounce
and exit rates. YouTube and Facebook have their own internal data collection methods. The
plethora of possible reports produced by products such GA, AWStats, Webalizer, YouTube, and
Facebook makes focusing on a select group of metrics important. For digital collections use,
most digital asset management systems have their own method to collect data. More important
than the method for digital collection use is a common understanding of the metrics’ definitions.
c. the priority that RBMS should place on developing a better metric and related
definitions and guidelines for use.
Google Analytics, and its associated metrics and definitions, offers a suitable tool for tracking
website use, though further comparison between GA and server log file products like AWStats
and Webalizer may be appropriate. RBMS could consider requesting that special and digital
collections units use a common tool in the event that differences between products make crossinstitutional comparison difficult. Regardless of the tool, RBMS should also consider how many
and which metrics are appropriate for assessment purposes. Given the extensive number of
“things” that can be counted in Google Analytics, the relative usefulness of each, and problems
associated with some measures, RBMS should decide what counts are most accurate,
informative, and important for assessment purposes. RBMS should also consider how much and
which data it wishes to report on digital collections use. Trends within ARL to decrease reporting
requirements on digital collections and ACRL’s continued use of more robust reporting should
be reconciled. In its efforts to improve assessment, RBMS should give priority to locating or
developing methodologies for documenting the impact and value of web page and digital
collection use.
Assessing Value and Impact for Users
Thomas Crichlow of Duke University Libraries offers several useful values, impacts, and
possible outcomes. The number of patron visits and page views in relative terms can help
institutions adjust how they design homepages and promote and support various resources.
Comparing on-campus and off-campus access to web pages or digital objects sheds light on
which audiences are served by the content. A better understanding of traffic sources can focus
RBMS Metrics Task Force, Subgroup 4
Website Usage / 5 of 5
greater institutional attention on access methods such as portals and search engine optimization.
Awareness of which browser platforms and devices patrons most often use can focus efforts on
developing effective web technologies and interfaces. Overall, determining patron demographics
and behavior for special and digital collections website and digital object use can inform the way
an institution marshals its resources and expends its energies. More difficult to ascertain is the
value these efforts have on patrons. Use does not equate with impact or value.
Documenting the impact and value of website and digital object use on patrons is far more
difficult than the statistical analysis that comes from products like Google Analytics. I am not
aware of any current national or international standards for determining impact, though I suspect
some institutions have done qualitative studies with their patrons and have sought to track the
use of digital objects in publications, etc.
A plethora of well-established metrics and tools exist for gathering quantitative measures to
assess website and digital collection use. These metrics are most suited for determining patron
demographics and behavior but not impact.
Based upon patron demographics and behavior, special and digital collections departments could
refocus financial and human resources away from content little used, build content to meet
proven demand, increase marketing efforts to grow demand, redesign web pages, and/or improve
tagging to aid in navigation and searchability.
Download