Effectively Applying Usage 
Statistics in E-Resource 
Collection Development 
Using Evidence and Outreach in Decision- 
Making 
ACRL-MD – New Identities: Adapting the Academic Library 
November 14, 2014 
Randy Lowe – Collection Development, Acquisition & Serials Librarian, Frostburg State University
Overview 
 Why E-Resources Assessment? 
 Usage Statistics – Types, Reports, 
Collection 
 Assessment: Evidence & Outreach 
◦ Applying usage statistics to collection 
management decision-making 
◦ Engaging librarians, faculty and 
administrators in the process
Why E-Resource 
Assessment? 
 Libraries have historically measured use 
of services (circulation statistics, re-shelving 
counts, gate counts, etc.) 
 The technology upon which e-resources 
reside inherently allows for extensive 
collection of usage data – and 
assessment of that use 
 Assessment of use data supports 
evidence-based collection management 
 Libraries operate in a challenging fiscal 
environment – demonstrating e-resource 
value and fiscal responsibility is a must
Effective E-Resources Assessment 
 Two essential elements in conducting 
effective e-resource assessments: 
◦ Efficient and Accurate Data Collection 
◦ Clear and Succinct Analysis 
 E-Resource assessment is more than 
just collecting usage statistics – it is 
applying them in the making of sound 
management decisions regarding 
library resources 
 Usage statistics measure volume, not 
value of resources
What Can You Do with E-Resources 
Usage Statistics? 
 Track usage / Assess overall collection use 
 Track expenditures / Figure cost-per-use 
 Track turnaways 
 Assess title, subject, publisher and other usage 
elements 
 Identify user behavior trends 
 Assist in making collection development 
decisions, including acquisition model selection 
 Effectively advocate for resources – especially 
if assessment is tied to institutional 
goals/strategic plan, curricular initiatives, 
student learning goals
Types of Usage Statistics Reports and 
When to Use Them 
 Vendor-Defined 
◦ Analyzing usage data from a single vendor 
◦ Obtaining cost information 
◦ Comprehensive data files make it easy to 
analyze combinations of various data 
elements [Example] 
◦ When COUNTER reports do not provide 
adequate detail 
 COUNTER-Compliant 
◦ Analyzing usage data across multiple 
vendors 
◦ Ensuring data integrity though adherence to 
recognized standards
Collecting Usage Data 
 Define Objectives 
◦ What you need to know or are trying to 
find out should drive your data collection 
decisions 
◦ Collecting Usage Statistics can be a major 
time commitment 
 Use your assessment objectives to help you to 
not only determine what data to collect, but 
when you have collected enough data to 
analyze 
 Properly balancing time and resources 
dedicated to both data collection and analysis 
is vital
Collecting Usage Data 
 Various vendors present data differently 
– this can present a challenge not only 
across vendors, but even with combining 
data elements from a single vendor 
 Manipulation / Formatting of raw data will 
likely be necessary 
 Example – COUNTER BR1 Report + 
Acquisition Type Data + Cost Data 
Compiled Manually = Data for 
Assessment 
 Schedule time(s) to collect data 
 Vendors’ archival policies for maintaining 
usage statistics vary
Assessing Usage Data 
You have usage data – What do you do 
with it? 
 It is easy to get overwhelmed in usage 
data – analysis should be guided by 
your assessment objectives 
◦ What do you want/need to assess? 
◦ What questions are you trying to answer? 
◦ Who is your audience? 
 Have a purpose for using your data
Assessing Usage Data 
 Assessment is most powerful when it 
is tied to an action or potential action 
(including requests) 
 There is no single method for 
assessing usage statistics in every 
case – the “right data” to analyze and 
include in your report is that which will 
support your assessment objectives
Usage Data Analysis 
 Data analysis should be thorough, but 
presented succinctly 
 Conclusions, trends, etc. should be 
clear and verifiable 
 Beware of pre-conceived notions, 
perceptions or opinions – hypotheses 
can be both proven and refuted 
 State known limitations of the data you 
have collected and how they may 
affect your analysis
Using/Applying Evidence: 
Writing Your Report 
 Know your audience 
 Include a brief purpose/introduction 
 Write clearly and succinctly 
 Reported usage data should support 
the purpose of the assessment 
◦ Only include data that supports your 
stated objectives – don’t include all 
collected data; it won’t be read by 
administrators
Using/Applying Evidence: 
Writing Your Report 
 Reported usage data should support the 
purpose of the assessment (continued) 
◦ Include data within the text of your report where it 
is necessary and provides clear evidence for the 
points you are making 
◦ It is usually more effective to include visual 
representations of (charts, graphs) rather than 
just figures within the text of reports 
◦ Larger tables and data sets, if necessary to 
include, are best placed in appendices 
 Conclusions and recommendations should be 
easily identified and based on the evidence 
presented 
 State action and/or desired response clearly
Using/Applying Evidence: 
The Frostburg Experience 
 Effectively applying e-resources data to 
collection management has been an 
evolution 
 The lay of the land – 2007 
◦ We had data (searches & link resolver) 
◦ Study to compare journal costs by format 
◦ Data sat in a vacuum outside of annual 
database budgeting 
 Needed to establish a frame of reference 
to begin applying usage statistics in 
engaging faculty and administrators
Evidence & Outreach Example 1: 
Faculty Survey – 2007-2008 
 Faculty had not been previously engaged 
systematically in collection development efforts 
 User behavior as demonstrated in link resolver 
statistics indicated that online full-text was 
preferred by users 
 Library determined periodicals and standing 
orders should be migrated to online format, but 
which ones? 
 Fall 2007: Faculty surveyed regarding value 
(content) and usefulness (format) of journals, 
standing orders, databases. 
 Spring 2008: Results of survey matched link 
resolver usage statistics 
 Subscription Cancellations, additions, format 
migrations made over next 5 years
Evidence & Outreach Example 2: 
Underutilized Journals 
 Library began collecting full text article 
retrievals in 2009-2010 (and re-shelving 
counts in 2011-2012) 
 All journal subscriptions are reviewed 
by librarians annually 
 Faculty are involved in second level of 
review for underutilized subscriptions 
 Objective is to use the process as a 
means for continued dialogue with 
faculty in collection development
Evidence & Outreach Example 3: 
Collaboration with Academic 
Depts  Academic departments becoming 
increasingly engaged in e-resource 
subscription discussions, including funding 
◦ Chemistry – CAS SciFinder 
◦ Visual Arts – Artstor 
 Current collaboration is with Biology 
◦ Department not satisfied with current e-resources 
◦ No funds available for additional resources 
◦ Reviewed use of current journal subscriptions 
and content of requested databases 
◦ Department suggested journal cancellations to 
fund databases 
◦ New e-resource scenarios developed
Evidence & Outreach Example 4: 
E-Book Assessment 
 Frostburg State University: Report 
overall use and expenditures of e-books 
over time; implement the most cost 
effective DDA acquisition model(s) 
[Report] 
 USMAI Consortial E-Book Pilot: Assess 
the effectiveness of a specific DDA 
acquisition model for the consortium; use 
and expenditures by consortium 
members and user types; identification of 
possible future program funding models 
[Report]
Thank You 
 Questions? 
 Contact Information: 
Randy Lowe 
Frostburg State University 
rlowe@frostburg.edu

Effectively Applying Usage Statistics in E-Resource Collection Development

  • 1.
    Effectively Applying Usage Statistics in E-Resource Collection Development Using Evidence and Outreach in Decision- Making ACRL-MD – New Identities: Adapting the Academic Library November 14, 2014 Randy Lowe – Collection Development, Acquisition & Serials Librarian, Frostburg State University
  • 2.
    Overview  WhyE-Resources Assessment?  Usage Statistics – Types, Reports, Collection  Assessment: Evidence & Outreach ◦ Applying usage statistics to collection management decision-making ◦ Engaging librarians, faculty and administrators in the process
  • 3.
    Why E-Resource Assessment?  Libraries have historically measured use of services (circulation statistics, re-shelving counts, gate counts, etc.)  The technology upon which e-resources reside inherently allows for extensive collection of usage data – and assessment of that use  Assessment of use data supports evidence-based collection management  Libraries operate in a challenging fiscal environment – demonstrating e-resource value and fiscal responsibility is a must
  • 4.
    Effective E-Resources Assessment  Two essential elements in conducting effective e-resource assessments: ◦ Efficient and Accurate Data Collection ◦ Clear and Succinct Analysis  E-Resource assessment is more than just collecting usage statistics – it is applying them in the making of sound management decisions regarding library resources  Usage statistics measure volume, not value of resources
  • 5.
    What Can YouDo with E-Resources Usage Statistics?  Track usage / Assess overall collection use  Track expenditures / Figure cost-per-use  Track turnaways  Assess title, subject, publisher and other usage elements  Identify user behavior trends  Assist in making collection development decisions, including acquisition model selection  Effectively advocate for resources – especially if assessment is tied to institutional goals/strategic plan, curricular initiatives, student learning goals
  • 6.
    Types of UsageStatistics Reports and When to Use Them  Vendor-Defined ◦ Analyzing usage data from a single vendor ◦ Obtaining cost information ◦ Comprehensive data files make it easy to analyze combinations of various data elements [Example] ◦ When COUNTER reports do not provide adequate detail  COUNTER-Compliant ◦ Analyzing usage data across multiple vendors ◦ Ensuring data integrity though adherence to recognized standards
  • 7.
    Collecting Usage Data  Define Objectives ◦ What you need to know or are trying to find out should drive your data collection decisions ◦ Collecting Usage Statistics can be a major time commitment  Use your assessment objectives to help you to not only determine what data to collect, but when you have collected enough data to analyze  Properly balancing time and resources dedicated to both data collection and analysis is vital
  • 8.
    Collecting Usage Data  Various vendors present data differently – this can present a challenge not only across vendors, but even with combining data elements from a single vendor  Manipulation / Formatting of raw data will likely be necessary  Example – COUNTER BR1 Report + Acquisition Type Data + Cost Data Compiled Manually = Data for Assessment  Schedule time(s) to collect data  Vendors’ archival policies for maintaining usage statistics vary
  • 9.
    Assessing Usage Data You have usage data – What do you do with it?  It is easy to get overwhelmed in usage data – analysis should be guided by your assessment objectives ◦ What do you want/need to assess? ◦ What questions are you trying to answer? ◦ Who is your audience?  Have a purpose for using your data
  • 10.
    Assessing Usage Data  Assessment is most powerful when it is tied to an action or potential action (including requests)  There is no single method for assessing usage statistics in every case – the “right data” to analyze and include in your report is that which will support your assessment objectives
  • 11.
    Usage Data Analysis  Data analysis should be thorough, but presented succinctly  Conclusions, trends, etc. should be clear and verifiable  Beware of pre-conceived notions, perceptions or opinions – hypotheses can be both proven and refuted  State known limitations of the data you have collected and how they may affect your analysis
  • 12.
    Using/Applying Evidence: WritingYour Report  Know your audience  Include a brief purpose/introduction  Write clearly and succinctly  Reported usage data should support the purpose of the assessment ◦ Only include data that supports your stated objectives – don’t include all collected data; it won’t be read by administrators
  • 13.
    Using/Applying Evidence: WritingYour Report  Reported usage data should support the purpose of the assessment (continued) ◦ Include data within the text of your report where it is necessary and provides clear evidence for the points you are making ◦ It is usually more effective to include visual representations of (charts, graphs) rather than just figures within the text of reports ◦ Larger tables and data sets, if necessary to include, are best placed in appendices  Conclusions and recommendations should be easily identified and based on the evidence presented  State action and/or desired response clearly
  • 14.
    Using/Applying Evidence: TheFrostburg Experience  Effectively applying e-resources data to collection management has been an evolution  The lay of the land – 2007 ◦ We had data (searches & link resolver) ◦ Study to compare journal costs by format ◦ Data sat in a vacuum outside of annual database budgeting  Needed to establish a frame of reference to begin applying usage statistics in engaging faculty and administrators
  • 15.
    Evidence & OutreachExample 1: Faculty Survey – 2007-2008  Faculty had not been previously engaged systematically in collection development efforts  User behavior as demonstrated in link resolver statistics indicated that online full-text was preferred by users  Library determined periodicals and standing orders should be migrated to online format, but which ones?  Fall 2007: Faculty surveyed regarding value (content) and usefulness (format) of journals, standing orders, databases.  Spring 2008: Results of survey matched link resolver usage statistics  Subscription Cancellations, additions, format migrations made over next 5 years
  • 16.
    Evidence & OutreachExample 2: Underutilized Journals  Library began collecting full text article retrievals in 2009-2010 (and re-shelving counts in 2011-2012)  All journal subscriptions are reviewed by librarians annually  Faculty are involved in second level of review for underutilized subscriptions  Objective is to use the process as a means for continued dialogue with faculty in collection development
  • 17.
    Evidence & OutreachExample 3: Collaboration with Academic Depts  Academic departments becoming increasingly engaged in e-resource subscription discussions, including funding ◦ Chemistry – CAS SciFinder ◦ Visual Arts – Artstor  Current collaboration is with Biology ◦ Department not satisfied with current e-resources ◦ No funds available for additional resources ◦ Reviewed use of current journal subscriptions and content of requested databases ◦ Department suggested journal cancellations to fund databases ◦ New e-resource scenarios developed
  • 18.
    Evidence & OutreachExample 4: E-Book Assessment  Frostburg State University: Report overall use and expenditures of e-books over time; implement the most cost effective DDA acquisition model(s) [Report]  USMAI Consortial E-Book Pilot: Assess the effectiveness of a specific DDA acquisition model for the consortium; use and expenditures by consortium members and user types; identification of possible future program funding models [Report]
  • 19.
    Thank You Questions?  Contact Information: Randy Lowe Frostburg State University [email protected]