giving SUSHI another try

(It's just) Kate's sushi! photo by Cindi Blyberg
photo by Cindi Blyberg

I’m going to give SUSHI another try this year. I had set it up for some of our stuff a few years back with mix results, so I removed it and have been continuing to manually retrieve and load reports into our consolidation tool. I’m still doing that for the 2017 reports, because the SUSHI harvesting tool I have won’t let me go back and pull from before, only monthly moving forward now.

I’ve spent a lot of time making sure titles in reports matched up with our ERMS so that consolidation would work (it’s matching on title, ugh), and despite my efforts, any reports generated still need cleanup. What is the value of my effort there? Not much anymore. Especially since ingesting cost data for journals/books is not a simple process to maintain, either. So, if all that matters less to none, might as well take whatever junk is passed along in the SUSHI feed as well and save myself some time for other work in 2019.

Charleston 2016: COUNTER Release 5 — Consistency, Clarity, Simplification and Continuous Maintenance

Speakers: Lorraine Estelle (Project COUNTER), Anne Osterman (VIVA – The Virtual Library of Virginia), Oliver Pesch (EBSCO Information Services)

COUNTER has had very minimal updates over the years, and it wasn’t until release 4 that things really exploded with report types and additional useful data. Release 5 attempts to reduce complexity so that all publishers and content providers are able to achieve compliance.

They are seeking consistency in the report layout, between formats, and in vocabulary. Clarity in metric types and qualifying action, processing rules, and formatting expectations.

The standard reports will be fewer, but more flexible. The expanded reports will introduce more data, but with flexibility.

A transaction will have different attributes recorded depending on the item type. They are also trying to get at intent — items investigated (abstract) vs. items requested (full-text). Searches will now distinguish between whether it was on a selected platform, a federated search, a discovery service search, or a search across a single vendor platform. Unfortunately, the latter data point will only be reported on the platform report, and still does not address teasing that out at the database level.

The access type attribute will indicate when the usage is on various Open Access or free content as well as licensed content. There will be a year of publication (YOP) attribution, which was not in any of the book reports and only included in Journal Report 5.

Consistent, standard header for each report, with additional details about the data. Consistent columns for each report. There will be multiple rows per title to cover all the combinations, making it more machine-friendly, but you can create filters in Excel to make it more human-friendly.

They expect to have release 5 published by July 2017 with compliance required by January 2019.

Q&A
Q: Will there eventually be a way to account for anomalies in data (abuse of access, etc.)?
A: They are looking at how to address use triggered by robot activity. Need to also be sensitive of privacy issues.

Q: Current book reports do not include zero use entitlements. Will that change?
A: Encouraged to provide KBART reports to get around that. The challenge is that DDA/PDA collections are huge and cumbersome to deliver reports. Will also be dropping the zero use reporting on journals, too.

Q: Using DOI as a unique identifier, but not consistently provided in reports. Any advocacy to include unique identifiers?
A: There is an initiative associated with KBART to make sure that data is shared so that knowledgbases are updated so that users find the content so that there are fewer zero use titles. Publisher have motivation to do this.

Q: How do you distinguish between unique uses?
A: Session based data. Assign a session ID to activity. If no session tracking, a combination of IP address and user agent. The user agent is helpful when multiple users are coming through one IP via the proxy server.

Slides

community site for usage statistics

Usus is an independent community website developed to help librarians, library consortium administrators, publishers, aggregators, etc. communicate around topics related to usage statistics. From problem-solving to workflow tips to calling out bad actors, this site hopes to be the hub of all things usage.

Do you have news to share or a problem you can’t figure out? Do you have really cool workflows you want to share? Drop us a note!

mapping ejournal use to subject areas

I had a thought last night as I was trying to fall asleep: what if I took our data on demand file that includes subjects and mashed it up with our consolidated JR1 use statistics? Could I get a better picture of the disciplines at my institution that are using ejournals? It’s definitely something worth looking at.

LibFest: Telling your Story with Usage Statistics — Making data work

presenter: Jamene Brooks-Kieffer

She won’t be talking about complex tools or telling you to hire more staff. Rather, she’ll be looking at ways we can use what we have to do it better.

Right now, we have too much data from too many sources, and we don’t have enough time or staff to deal with it. And, nobody cares about it anyway. Instead of feeling blue about this, change your attitude.

Start by looking at smaller chunks. Look at all of the data types and sources, then choose one to focus on. Don’t stress about the rest. How to pick which one? Select data that has been consistently collected over time. If it’s focused on a specific activity, it’ll be easier to create a story about it. And finally, the data should be both interesting and accessible to you.

By selecting only one source of data, you have reduced the stress on time. You also need to acknowledge your limits in order to move forward. You can’t work miracles, but you can show enough impact to get others on board. Tie the data to your organizational goals. Analyze the data using the tools you already have (i.e. Excel), and then publicize the results of your work.

Why use Excel? It’s pretty universal, and there are free alternatives for spreadsheets if you need them. Three useful Excel tools: import & manipulate files of various formats (CSV files), consolidate similar information (total annual data from monthly worksheets), and conditional formatting (identify cost/use over thresholds).

The spreadsheets are for you, not the stakeholders. Stop relying on them to communicate your data. The trouble with spreadsheets is that although they contain a lot of data, it’s challenging for those unfamiliar with the sources to understand the meaning of the data. Sending a summary/story will get your message across faster and more clearly.

Data has context, settings, complexities, and conflicts. One of the best ways of communicating it is through a story. Give stakeholders the context to hang the numbers on and a way to remember why they are important. Write what you know, focus on the important things, and keep it brief and meaningful. Here is an example: Data Stories: A dirty job.

Data stories are everywhere. It’s not strictly for usage or financial data. If you have a specific question you want answered through data, it makes it easier to compose the story.

Convince yourself to act; your actions will persuade others.

presenter: Katy Silberger

She will be showing three scenarios for observing user behavior through statistics: looking at the past with vendor supplied statistics, assessing current user behavior with Google Analytics, and anticipating user behavior with Google Analytics.

They started looking at usage patterns before and after implementing federated searching. It was hard to answer the question of how federated searching changed user behavior. They used vendor usage reports and website visits to calculate the number of articles retrieved per website visit and articles retrieved per search. They found that the federated search tool generated an increase in article/use. The ratios take into account the fluctuation in user populations.

Google Analytics could be used to identify use from students abroad. It’s also helpful for identifying trends in mobile web access.