IL 2012: Discovery Systems

Space Shuttle Discovery Landing At Washington DC
“Space Shuttle Discovery Landing At Washington DC” by Glyn Lowe

Speaker: Bob Fernekes

The Gang of Four: Google, Apple, Amazon, & Facebook

Google tends to acquire companies to grow the capabilities of it. We all know about Apple. Amazon sells more ebooks than print books now. Facebook is… yeah. That.

And then we jump to selecting a discovery service. You would do that in order to make the best use of the licensed content. This guy’s library did a soft launch in the past year of the discovery service they chose, and it’s had an impact on the instruction and tools (i.e. search boxes) he uses.

And I kind of lost track of what he was talking about, in part because he jumped from one thing to the next, without much of a transition or connection. I think there was something about usability studies after they implemented it, although they seemed to focus on more than just the discovery service.

Speaker: Alison Steinberg Gurganus

Why choose a discovery system? You probably already know. Students lack search skills, but they know how to search, so we need to give them something that will help them navigate the proprietary stuff we offer out on the web.

The problem with the discovery systems is that they are very proprietary. They don’t quite play fairly or nicely with competitor’s content yet.

Our users need to be able to evaluate, but they also need to find the stuff in the first place. A great discovery service should be self-explanatory, but we don’t have that yet.

We have students who understand Google, which connects them to all the information and media they want. We need something like that for our library resources.

When they were implementing the discovery tool, they wanted to make incremental changes to the website to direct users to it. They went from two columns, with the left column being text links to categories of library resources and services, to three columns, with the discover search box in the middle column.

When they were customizing the look of the discovery search results, they changed the titles of items to red (from blue). She notes that users tend to ignore the outside columns because that’s where Google puts advertisements, so they are looking at ways to make that information more visible.

I also get the impression that she doesn’t really understand how a discovery service works or what it’s supposed to do.

Speaker: Athena Hoeppner

Hypothesis: discovery includes sufficient content of high enough quality, with full text, and …. (didn’t type fast enough).

Looked at final papers from a PhD level course (34), specifically the methodology section and bibliography. Searched for each item in the discovery search as well as one general aggregator database and two subject-specific databases. The works cited were predominately articles, with a significant number of web sources that were not available through library resources. She was able to find more citations in the discovery search than in Google Scholar or any of the other library databases.

Clearly the discovery search was sufficient for finding the content they needed. Then they used a satisfaction survey of the same students that covered familiarity and frequency of use for the subject indexes, discovery search, and Google Scholar. Ultimately, it came down that the students were satisfied and happy with the subject indexes, and too few respondents to get a sense of satisfaction with the discovery search or Google Scholar.

Conclusions: Students are unfamiliar with the discovery system, but it could support their research needs. However, we don’t know if they can find the things they are looking for in it (search skills), nor do we know if they will ultimately be happy with it.

NASIG 2012: A Model for Electronic Resources Assessment

Presenter: Sarah Sutton, Texas A&M University-Corpus Christi

Began the model with the trigger event — a resource comes up for renewal. Then she began looking at what information is needed to make the decision.

For A&I databases, the primary data pieces are the searches and sessions from the COUNTER release 3 reports. For full-text resources, the primary data pieces are the full-text downloads also from the COUNTER reports. In addition to COUNTER and other publisher supplied usage data, she looks at local data points. Link-outs from the a-to-z list of databases tells her what resources her users are consciously choosing to use, and not necessarily something they arrive at via a discovery service or Google. She’s able to pull this from the content management system they use.

Once the data has been collected, it can be compared to the baseline. She created a spreadsheet listing all of the resources, with a column each for searches, sessions, downloads, and link-outs. The baseline set of core resources was based on a combination of high link-outs and high usage. These were grouped by similar numbers/type of resource. Next, she calculated the cost/use for each of the four use types, as well as the percentage of change in use over time.

After the baseline is established, she compares the renewing resource to that baseline. This isn’t always a yes or no answer, but more of a yes or maybe answer. Often more analysis is needed if it is tending towards no. More data may include overlap analysis (unique to your library collection), citation lists (unique titles — compare them with a list of highly-cited journals at your institution or faculty requests or appear on a core title list), journal-level usage of the unique titles, and impact factors of the unique titles.

Audience question: What about qualitative data? Talk to your users. Does not have a suggestion for how to incorporate that into the model without increasing the length of time in the review process.

Audience question: How much staff time does this take? Most of the work is in setting up the baseline. The rest depends on how much additional investigation is needed.

[I had several conversations with folks after this session who expressed concern with the method used for determining the baseline. Namely, that it excludes A&I resources and assumes that usage data is accurate. I would caution anyone from wholesale adopting this as the only method of determining renewals. Without conversation and relationships with faculty/departments, we may not truly understand what the numbers are telling us.]

NASIG 2012: Mobile Websites and APP’s in Academic Libraries Harmony on a Small Scale

Speaker: Kathryn Johns-Masten, State University of New York Oswego

About half of American adults have smart phones now. Readers of e-books tend to read more frequently than others. They may not be reading more academic material, but they are out there reading.

SUNY Oswego hasn’t implemented a mobile site, but the library really wanted one, so they’ve created their own using the iWebKit from MIT.

Once they began the process of creating the site, they had many conversations about who they were targeting and what they expected to be used in a mobile setting. They were very selective about which resources were included, and considered how functional each tool was in that setting. They ended up with library hours, contact, mobile databases, catalog, ILL article retrieval (ILLiad), ask a librarian, Facebook, and Twitter (in that order).

When developing a mobile site, start small and enhance as you see the need. Test functionality (pull together users of all types of devices at the same time, because one fix might break another), review your usage statistics, and talk to your users. Tell your users that it’s there!

Tools for designing your mobile site: MobiReady, Squeezer, Google Mobile Site Builder, Springshare Mobile Site Builder, Boopsie, Zinadoo, iWebKit, etc.

Other things related to library mobile access… Foursquare! The library has a cheat sheet for answers to the things freshman are required to find on campus, so maybe they could use Foursquare to help with this. Tula Rosa Public Library used a screen capture of Google Maps to help users find their new location. QR codes could link to ask a librarian, book displays linked to reviews, social media, events, scavenger hunts, etc. Could use them to link sheet music to streaming recordings.

NASIG 2012: Managing E-Publishing — Perfect Harmony for Serialists

Presenters: Char Simser (Kansas State University) & Wendy Robertson (University of Iowa)

Iowa looks at e-publishing as an extension of the central mission of the library. This covers not only text, but also multimedia content. After many years of ad-hoc work, they formed a department to be more comprehensive and intentional.

Kansas really didn’t do much with this until they had a strategic plan that included establishing an open access press (New Prairie). This also involved reorganizing personnel to create a new department to manage the process, which includes the institutional depository. The press includes not only their own publications, but also hosts publications from a few other sources.

Iowa went with BEPress’ Digital Commons to provide both the repository and the journal hosting. Part of why they went this route for their journals was because they already had it for their repository, and they approach it more as being a hosting platform than as being a press/publisher. This means they did not need to add staff to support it, although they did add responsibilities to exiting staff in addition to their other work.

Kansas is using Open Journal Systems hosted on a commercial server due to internal politics that prevented it from being hosted on the university server. All of their publications are Gold OA, and the university/library is paying all of the costs (~$1700/year, not including the .6 FTE staff hours).

Day in the life of New Prairie Press — most of the routine stuff at Kansas involves processing DOI information for articles and works-cited, and working with DOAJ for article metadata. The rest is less routine, usually involving journal setups, training, consultation, meetings, documentation, troubleshooting, etc.

The admin back-end of OJS allows Char to view it as if she is different types of users (editor, author, etc.) to be able to trouble-shoot issues for users. Rather than maintaining a test site, they have a “hidden” journal on the live site that they use to test functions.

A big part of her daily work is submitting DOIs to CrossRef and going through the backfile of previously published content to identify and add DOIs to the works-cited. The process is very manual, and the error rate is high enough that automation would be challenging.

Iowa does have some subscription-based titles, so part of the management involves keeping up with a subscriber list and IP addresses. All of the titles eventually fall into open access.

Most of the work at Iowa has been with retrospective content — taking past print publications and digitizing them. They are also concerned with making sure the content follows current standards that are used by both library systems and Google Scholar.

There is more. I couldn’t take notes and keep time towards the end.

NASIG 2012: Why the Internet is More Attractive Than the Library

Speaker: Dr. Lynn Silipigni Connaway, OCLC

Students, particularly undergraduates, find Google search results to make more sense than library database search results. In the past, these kinds of users had to work around our services, but now we need to make our resources fit their workflow.

Connaway has tried to compare 12 different user behavior studies in the UK and the US to draw some broad conclusions, and this has informed her talk today.

Convenience is number one, and it changes. Context and situation are very important, and we need to remember that when asking questions about our users. Sometimes they just want the answer, not instruction on how to do the research.

Most people power browse these days: scan small chunks of information, view first few pages, no real reading. They combine this with squirreling — short, basic searches and saving the content for later use.

Students prefer keyword searches. This is supported by looking at the kinds of terms used in the search. Experts use broad terms to cover all possible indexing, novices use specific terms. So why do we keep trying to get them to use the “advance” search in our resources?

Students are confident with information discovery tools. They mainly use their common sense for determining the credibility of a site. If a site appears to have put some time into the presentation, then they are more likely to believe it.

Students are frustrated with navigating library websites, the inconvenience of communicating with librarians face to face, and they tend to associate libraries only with books, not with other information. They don’t recognize that the library is who is providing them with access to online content like JSTOR and the things they find in Google Scholar.

Students and faculty often don’t realize they can ask a question of a librarian in person because we look “busy” staring at our screens at the desk.

Researchers don’t understand copyright, or what they have signed away. They tend to be self-taught in discovery, picking up the same patterns as their graduate professors. Sometimes they rely on the students to tell them about newer ways of finding information.

Researchers get frustrated with the lack of access to electronic backfiles of journals, discovering non-English content, and unavailable content in search results (dead links, access limitation). Humanities researchers feel like there is a lack of good, specialized search engines for them (mostly for science). They get frustrated when they go to the library because of poor usability (i.e. signs) and a lack of integration between resources.

Access is more important than discovery. They want a seamless transition from discovery to access, without a bunch of authentication barriers.

We should be improving our OPACs. Take a look at Trove and Westerville Public Library. We need to think more like startups.

tl;dr – everything you’ve heard or read about what our users really do and really need, but we still haven’t addressed in the tools and services we offer to them

ER&L 2012: Taking the Guesswork Out of Demand-Driven Acquisition — Two Approaches

Tome Reader
photo by QQ Li

Speakers: Carol J. Cramer & Derrik Hiatt

They did an analysis of their circulating print collection to see what areas or books would have the equivalent uses to trigger a purchase if it were electronic. Only 2% of their entire circulating collection met the trigger point to where it would be more cost effective to purchase than to go with a short term loan option.

They announced the DDA trial, but deliberately did not tell the users that it would incur cost, just that it was there. They would pay short term loans up to the sixth use, and then they would purchase the title. The year of usage gave them an idea of what adjustments needed to be made to the trigger point. Eventually, the cost flattens out at the sixth use, and the difference between continuing to pay STLs and buying the book is small.

They were able to identify if the triggered purchase book was used by a single person (repeatedly), by a class (several people), or a mix of both, and it was split in almost even thirds.

They determined that 6 was a good trigger. The STL cost ended up being an average of 10.5% of the list cost. DDA doesn’t have to break the bank, and was lower than expected. The number of titles in the catalog didn’t have as much to do with the amount spent as the FTE. It also lead to questioning the value of firm ordering ebooks rather than letting DDA cover it

However, this is only 11 months of data, and more longitudinal studies are needed.

Speaker: Lea Currie

They loaded records for slip books, and then the users have the option to request them at various levels of speed. The users are notified when the print book arrives, and the full MARC record is not loaded until the book is returned.

They saved quit a bit of money per month using this method, and 88% of the titles purchased circulated. Only about 75% of their ILL titles will circulate, to put that into perspective.

Of course, librarians still had some concerns. First, the library catalog is not an adequate tool for discovering titles. Faculty were concerned about individuals doing massive requests for personal research topics. Also, faculty do not want to be selectors for the libraries. [ORLY? They want the books they want when they want them — how is that different?]

The next DDA project was for ebooks, using the typical trigger points. They convinced the Social Science and Sci/Tech librarians to put a price cap for DDA titles. Up to a certain price, the book would be included in the approval plan, between a range it would go in DDA, and then above that range it would require the librarian’s approval. These were written into their YBP profile.

For the pDDA, they discovered that as the books aged, it was harder to do rush orders since they were going out of print. They also modified their language to indicate that the books may not be available if they are out of print.

They have not done DDA for humanities or area studies. They based their decisions on the YBP profile on retrospective reports, which allowed them to get an idea of the average cost.

For FY12, they expect that the breakdown will be 23% eDDA, 50% pDDA, 20% approval, and 7% selected by subject bibliographers. They’ve also given the subject librarians the options to review the automatic approval ebooks — they have a week to reject or shift to DDA each title if they want. They can also shift the expensive titles to DDA if they want to see if anyone would use it before choosing to purchase it.

Questions:
Are you putting the records in your discovery service if you have one, and can you tell if the uses are coming from that or your catalog? Not yet. Implementing a discovery service. Some find resources through Google Scholar.

musings on web-scale discovery systems

photo by Pascal

My library is often on the forefront of innovation, having the advantage of a healthy budget and staff size, yet small enough to be nimble. Frequently, when my colleagues return from conferences and give their reports, they’ll conclude with something along the lines of “we’re already doing most of the things they talked about.” At a recent conference report session, that was repeated again, with one exception: we have not implemented a web-scale discovery system.

I’m of two minds about web-scale discovery systems. In theory, they’re pretty awesome, allowing users to discover all of the content available to them from the library, regardless of the source or format. But in reality, they’re hamstrung by exclusive deals and coding limitations. The initial buzz was that they caused a dramatic increase in the use of library resources, but a few years in, and I’m hearing conflicting reports and grumblings.

We held off on buying a web-scale discovery system for two main reasons: one, we didn’t have the funding secured, and two, most of the reference librarians felt indifferent to outright dislike towards the systems out there at the time. We’re now in the process of reviewing and evaluating the current systems available, after many discussions about which problems we are hoping they will solve.

In the end, they really aren’t “Google for Libraries.” We think that our users want a single search box, but do they really? I heard an anecdote about how the library had spent a lot of time teaching users where to find their web-scale discovery system, making sure it was visible on the main library page, etc. After a professor assigned the same students to find a known article (gave them the full citation) using the web-scale discovery system (called it by name), the most frequent question the library got was, “How do I google the <name of web-scale discovery system>?”

I wonder if the ROI really is significant enough to implement and promote a web-scale discovery system? These systems are not cheap, and they take a bit of labor to maintain them. And, frankly, if the battle over exclusive content continues to be waged, it won’t be easy to pick the best one for our collection/users and know that it will stay that way for more than six months or a year.

Does your library have a web-scale discovery system? Is it everything you thought it would be? Would you pick the same one if you had to choose again?

css.php