Ithaka’s What to Withdraw tool

Have you seen the tool that Ithaka developed to determine what print scholarly journals you could withdraw (discard/store) that are already in your digital collections? It’s pretty nifty for a spreadsheet. About 10-15 minutes of playing with it and a list of our print holdings resulted in giving me a list of around 200 or so actionable titles in our collection, which I passed on to our subject liaison librarians.

The guys who designed it are giving some webinar sessions, and I just attended one. Here are my notes, for what it’s worth. I suggest you participate in a webinar if you’re interested in it. The next one is tomorrow and there’s one on February 10th as well.


Background

  • They have an organizational commitment to preservation: JSTOR, Portico, and Ithaka S+R
  • Libraries are under pressure to both decrease their print collections and to maintain some print copies for the library community as a whole
  • Individual libraries are often unable to identify materials that are sufficiently well-preserved elsewhere
  • The What to Withdraw framework is for general collections of scholarly journals, not monographs, rare books, newspapers, etc.
  • The report/framework is not meant to replace the local decision-making process

What to Withdraw Framework

  • Why do we need to preserve the print materials once we have a digital version?
    • Fix errors in the digital versions
    • Replace poor quality scans or formats
    • Inadequate preservation of the digital content
    • Unreliable access to the digital content
    • Also, local politics or research needs might require access to or preservation of the print
  • Once they developed the rationales, they created specific preservation goals for each category of preservation and then determined the level of preservation needed for each goal.
    • Importance of images in journals (the digitization standards for text is not the same as for images, particularly color images)
    • Quality of the digitization process
    • Ongoing quality assurance processes to fix errors
    • Reliability of digital access (business model, terms & conditions)
    • Digital preservation
  • Commissioned Candace Yano (operations researcher at UC Berkeley) to develop a model for copies needed to meet preservation goals, with the annual loss rate of 0.1% for a dark archive.
    • As a result, they found they needed only two copies to have a >99% confidence than they will still have remaining copies left in twenty years.
    • As a community, this means we need to be retaining at least two copies, if not more.

Decision-Support Tool (proof of concept)

  • JSTOR is an easy first step because many libraries have this resource and many own print copies of the titles in the collections and Harvard & UC already have dim/dark archives of JSTOR titles
  • The tool provides libraries information to identify titles held by Harvard & UC libraries which also have relatively few images

Future Plans

  • Would like to apply the tool to other digital collections and dark/dim archives, and they are looking for partners in this
  • Would also like to incorporate information from other JSTOR repositories (such as Orbis-Cascade)

IL2009: Mashups for Library Data

Speakers: Nicole Engard

Mashups are easy ways to provide better services for our patrons. They add value to our websites and catalogs. They promote our services in the places our patrons frequent. And, it’s a learning experience.

We need to ask our vendors for APIs. We’re putting data into our systems, so we should be able to get it out. Take that data and mash it up with popular web services using RSS feeds.

Yahoo Pipes allows you to pull in many sources of data and mix it up to create something new with a clean, flow chart like interface. Don’t give up after your first try. Jody Fagan wrote an article in Computers in Libraries that inspired Engard to go back and try again.

Reading Radar takes the NYT Bestseller lists and merges it with data from Amazon to display more than just sales information (ratings, summaries, etc.). You could do that, but instead of having users go buy the book, link it to your library catalog. The New York Times has opened up a tremendous amount of content via APIs.

Bike Tours in CA is a mashup of Google Maps and ride data. Trulia, Zillow, and HousingMaps use a variety of sources to map real estate information. This We Know pulls in all sorts of government data about a location. Find more mashups at ProgrammableWeb.

What mashups should libraries be doing? First off, if you have multiple branches, create a Google Maps mashup of library locations. Share images of your collection on Flickr and pull that into your website (see Access Ceramics), letting Flickr do the heavy lifting of resizing the images and pulling content out via machine tags. Delicious provides many options for creating dynamically updating lists with code snippets to embed them in your website.

OPAC mashups require APIs, preferably those that can generate JavaScript, and finally you’ll need a programmer if you can’t get the information out in a way you can easily use it. LexisNexis Academic, WorldCat, and LibraryThing all have APIs you can use.

Ideas from Librarians: Mashup travel data from circulation data and various travel sources to provide better patron services. Grab MARC location data to plot information on a map. Pull data about media collection and combine it with IMDB and other resources. Subject RSS feeds from all resources for current articles (could do that already with a collection of journals with RSS feeds and Yahoo Pipes).

Links and more at her book website.

NASIG 2009: Informing Licensing Stakeholders

Towards a More Effective Negotiation

Presenters: Lisa Sibert, Micheline Westfall, Selden Lamoreux, Clint Chamberlain (moderator), Vida Damijonaitis, and Brett Rubinstein

Licensing as a process has not been improving very much. Some publishers are willing to negotiate changes, but some are still resistant. It often takes months to a year to receive fully signed licenses from publishers, which can tie up access or institutional processes. Negotiation time is, of course, a factor, but it should not effect the time it takes for both parties to sign and distribute copies once the language is agreed upon. One panelist noted that larger publishers are often less willing to negotiate than smaller ones. Damijonaitis stated that licenses are touched at fourteen different points in the process on their end, which plays into the length of time.

Publishers are concerned with the way the content is being used and making sure that it is not abused (without consequences). Is it necessary to put copyright violation language in licenses or can it live on purchase orders? Springer has not had any copyright violations that needed to be enforced in the past five or six years. They work with their customers to solve any problems as they come up, and libraries have been quick to deal with the situation. On the library side, some legal departments are not willing to allow libraries to participate in SERU.

Deal breakers: not allowing walk-ins, adjunct faculty, interlibrary loan, governing law, and basic fair use provisions. Usage statistics and uptime guarantees are important and sometimes difficult to negotiate. LibLicense is useful for getting effective language that publishers have agreed to in the past.

It’s not the libraries who tend to be the abusers of license terms or copyright, it’s the users. Libraries are willing to work with publishers, but if the technology has grown to the point where it is too difficult for the library to police use, then some other approach is needed. When we work with publishers that don’t require licenses or use just purchase orders, there is less paperwork, but it also doesn’t indemnify the institution, which is critical in some cases.

Bob Boissy notes that no sales person gets any benefit from long negotiations. They want a sale. They want an invoice. Libraries are interested in getting the content as quickly as possible. I think we all are coming at this with the same desired outcome.

NASIG 2009: ERMS Integration Strategies – Opportunity, Challenge, or Promise?

Speakers: Bob McQuillan (moderator), Karl Maria Fattig, Christine Stamison, and Rebecca Kemp

Many people have an ERM, some are implementing it, but few (in the room) are where they consider to be finished. ERMS present new opportunity and challenges with workflow and staffing, and the presenters intend to provide some insight for those in attendance.

At Fattig’s library, their budget for electronic is increasing as print is decreasing, and they are also running out of space for their physical collections. Their institution’s administration is not supportive of increasing space for materials, so they need to start thinking about how to stall or shrink their physical collection. In addition, they have had reductions in technical services staffing. Sound familiar?

At Kemp’s library, she notes that about 40% of her time is spent on access setup and troubleshooting, which is an indication of how much of their resources is allocated for electronic resources. Is it worth it? They know that many of their online resources are heavily used. Consorital “buying clubs” makes big deals possible, opening up access to more resources than they could afford on their own. Electronic is a good alternative to adding more volumes to already over-loaded shelves.

Stamison (SWETS) notes that they have seen a dramatic shift from print to electronic. At least two-thirds of the subscriptions they handle have an electronic component, and most libraries are going e-only when possible. Libraries tell them that they want their shelf space. Also, many libraries are going direct to publishers for the big deals, with agents getting involved only for EDI invoicing (cutting into the agent’s income). Agents are now investing in new technologies to assist libraries in managing e-collections, including implementing access.

Kemp’s library had a team of three to implement Innovative’s ERM. It took a change in workflow and incorporating additional tasks with existing positions, but everyone pulled through. Like libraries, Stamison notes that agents have had to change their workflow to handle electronic media, including extensive training. And, as libraries have more people working with all formats of serials, agents now have many different contacts within both libraries and publishers.

Fattig’s library also reorganized some positions. The systems librarian, acquisitions librarian, and serials & electronic resources coordinator all work with the ERMS, pulling from the Serials Solutions knowledge base. They have also contracted with someone in Oregon to manage their EZproxy database and WebBridge coverage load. Fattig notes that it takes a village to maintain an ERMS.

Agents with electronic gateway systems are working to become COUNTER compliant, and are heavily involved with developing SUSHI. Some are also providing services to gather those statistics for libraries.

Fattig comments that usage statistics are serials in themselves. At his library, they maintained a homegrown system for collecting usage statistics from 2000-07, then tried Serials Solutions Counter 360 for a year, but now are using an ERM/homegrown hybrid. They created their own script to clean up the files, because as we all know, COUNTER compliance means something different to each publisher. Fattig thinks that database searches are their most important statistics for evaluating platforms. They use their federated search statistics to weigh the statistics from those resources (will be broken out in COUNTER 3 compliance).

Kemp has not been able to import their use stats into ERM. One of their staff members goes in every month to download stats, and the rest come from ScholarlyStats. They are learning to make XML files out of their Excel files and hope to use the cost per use functionality in the future.

Fattig: “We haven’t gotten SUSHI to work in some of the places it’s supposed to.” Todd Carpenter from NISO notes that SUSHI compliance is a requirement of COUNTER 3.

For the next 12-18 months, Fattig expects that they will complete the creation of license and contact records, import all usage data, and implement SUSHI when they can. They will continue to work with their consorital tool, implement a discovery layer, and document everything. Plans to create a “cancellation ray gun and singalong blog” — a tool for taking criteria to generate suggested cancellation reports.

Like Fattig, Kemp plans to finish loading all of the data about license and contacts, also the coverage data. Looking forward to eliminating a legacy spreadsheet. Then, they hope to import COUNTER stats and run cost/use reports.

Agents are working with ONIX-PL to assist libraries in populating their ERMS with license terms. They are also working with CORE to assist libraries with populating acquisitions data. Stamison notes that agents are working to continue to be liaisons between publishers, libraries, and system vendors.

Dan Tonkery notes that he’s been listening to these conversations for years. No one is serving libraries very well. Libraries are working harder to get these things implemented, while also maintaining legacy systems and workarounds. “It’s too much work for something that should be simple.” Char Simser notes that we need to convince our administrations to move more staff into managing eresources as our budgets are shifting more towards them.

Another audience member notes that his main frustration is the lack of cooperation between vendors/products. We need a shared knowledge base like we have a shared repository for our catalog records. This gets tricky with different package holdings and license terms.

Audience question: When will the ERM become integrated into the ILS? Response: System vendors are listening, and the development cycle is dependent on customer input. Every library approaches their record keeping in different ways.

NASIG 2009: Ensuring Perpetual Access to Online Subscriptions

Presenters: Judy Luther (moderator), Ken DiFiore, Nancy Gibbs (contributed, but not able to attend), Selden Lamorueux, Victoria Reich, Heather Ruland Staines, and Kim Steinle

Librarians have asked publishers to ensure that they will have perpetual access available to paid subscribers, which is a reverse of the previous arrangement with print, which required the librarians to preserve access to the physical copy in their buildings. It took publishers some time to address this, and it continues to evolve.

Libraries, and academic libraries in particular, have proven staying power. Librarians, working with publishers, are responsible for providing access to information.

There are two issues with online perpetual access: maintaining access to content when a subscription ends and maintaining access to content when a publisher ceases to exist. Libraries have the resources to be custodians of the content through digital preservation practices, both alone and in community.

How are librarians & publishers managing expectations for the financial impact of moving to online-only content? First, make sure that there is some sort of preservation guarantee. You will need to start shifting staff from managing stacks/materials to managing online, which will impact training needs. There are cost savings in space and opening up the building to other relevant uses. Peter McCracken suggests that we should emphasize the service and benefits of online content.

Libraries purchase the delivery mechanism, not the content. Owning a print archive does not mean that online is automatically free. And, it’s not just the content that you get with online. The user interface is an essential component, and it isn’t cheap to develop or maintain.

The most important thing that librarians need to do is read the license agreement. If it doesn’t specify what happens to content if a subscription is canceled or if the publisher ceases to exist, negotiate for that to be added. CYA, folks! If you pay for it, make sure you’re really going to get it.

NASIG 2009: Ambient Findability

Libraries, Serials, and the Internet of Things

Presenter: Peter Morville

He’s a librarian that fell in love with the web and moved into working with information architecture. When he first wrote the book Information Architecture, he and his co-author didn’t include a definition of information architecture. With the second edition, they had four definitions: the structural design of shared information environments; the combination of organization, labeling, search, and navigation systems in webs sites and intranet; the art and science of shaping information products and experiences to support usability and finadability; an emerging discipline and community of practice focused on bringing principles of designing and architecture to the digital landscape.

[at this point, my computer crashed, losing all the lovely notes I had taken so far]

Information systems need to use a combination of categories (paying attention to audience and taxonomy), in-text linking, and alphabetical indexes in order to make information findable. We need to start thinking about the information systems of the future. If we examine the trends through findability, we might have a different perspective. What are all the different ways someone might find ____? How do we describe it to make it more findable?

We are drowning in information. We are suffering from information anxiety. Nobel Laureate Economist Herbert Simon said, “A wealth of information creates a poverty of attention.”

Ambient devices are alternate interfaces that bring information to our attention, and Moreville thinks this is a direction that our information systems are moving towards. What can we now do when our devices know where we are? Now that we can do it, how do we want to use it, and in what contexts?

What are our high-value objects, and is it important to make them more findable? RFID can be used to track important but easily hidden physical items, such as wheelchairs in a hospital. What else can we do with it besides inventory books?

In a world where for every object there are thousands of similar objects, how do we describe the uniqueness of each one? Who’s going to do it? Not Microsoft, and not Donald Rumsfeld and his unknown unknown. It’s librarians, of course. Nowadays, metadata is everywhere, turning everyone who creates it into librarians and information architects of sorts.

One of the challenges we have is determine what aspects of our information systems can evolve quickly and what aspects need more time.

In five to ten years from now, we’ll still be starting by entering a keyword or two into a box and hitting “go.” This model is ubiquitous and it works because it acknowledges human psychology of just wanting to get started. Search is not just about the software. It’s a complex, adaptive system that requires us to understand our users so that they not only get started, but also know how to take the next step once they get there.

Some example of best and worse practices for search are on his Flickr. Some user-suggested improvements to search are auto-compete search terms, suggested links or best bets, and for libraries, federated search helps users know where to begin. Faceted navigation goes hand in hand with federated search, which allows users to formulate what in the past would have been very sophisticated Boolean queries. It also helps them to understand the information space they are in by presenting a visual representation of the subset of information.

Morville referenced last year’s presentation by Mike Kuniavsky regarding ubiquitous computing, and he hoped that his presentation has complemented what Kuniavsky had to say.

Libraries are more than just warehouses of materials — they are cathedrals of information that inspire us.

PDF of his slides

CIL 2009: ERM… What Do You Do With All That Data, Anyway?

This is the session that I co-presented with Cindi Trainor (Eastern Kentucky University). The slides don’t convey all of the points we were trying to make, so I’ve also included a cleaned-up version of those notes.

  1. Title
  2. In 2004, the Digital Library Federation (DLF) Electronic Resources Management Initiative (ERMI) published their report on the electronic resource management needs of libraries, and provided some guidelines for what data needed to be collected in future systems and how that data might be organized. The report identifies over 340 data elements, ranging from acquisitions to access to assessment.

    Libraries that have implemented commercial electronic resource management systems (ERMS) have spent many staff hours entering data from old storage systems, or recording those data for the first time, and few, if any, have filled out each data element listed in the report. But that is reasonable, since not every resource will have relevant data attached to it that would need to be captured in an ERMS.

    However, since most libraries do not have an infinite number of staff to focus on this level of data entry, the emphasis should instead be placed upon capturing data that is neccessary for managing the resources as well as information that will enhance the user experience.

  3. On the staff side, ERM data is useful for: upcoming renewal notifications, generating collection development reports that explain cost-per-use, based on publisher-provided use statistics and library-maintained, acquisitions data, managing trials, noting Electronic ILL & Reserves rights, and tracking the uptime & downtime of resources.
  4. Most libraries already have access management systems (link resolvers, A-Z lists, Marc records).
  5. User issues have shifted from the multiple copy problem to a “which copy?” problem. Users have multiple points of access, including: journal packages (JSTOR, Muse); A&I databases, with and without FT (which constitute e-resources in themselves); Library website (particularly “Electronic Resources” or “Databases” lists); OPAC; A-Z List (typically populated by an OpenURL link resolver); Google/gScholar; article/paper references/footnotes; course reserves; course management systems (Blackboard, Moodle, WebCT, Angel,Sakai); citation management software (RefWorks, EndNote, Zotero); LibGuides / course guides; bookmarks
  6. Users want…
  7. Google
  8. Worlds collide! What elements from the DLF ERM spec could enhance the user experience, and how? Information inside an ERMS can enhance access management systems or discovery: subject categorization within the ERM that would group similar resources and allow them to be presented alongside the resource that someone is using; using statuses to group & display items, such as a trialset within the ERM to automatically populate a page of new resources or an RSS feed to make it easy for the library to group and publicize even 30 day trial. ERMS’s need to do a better job of helping to manage the resource lifecycle by being built to track resources through that lifecycle so that discovery is updated by extension because resources are managed well, increasing uptime and availability and decreasing the time from identification above potential new resource to accessibility of that resource to our users
  9. How about turning ERM data into a discovery tool? Information about accessibility of resources to reference management systems like Endnote, RefWorks, or Zotero, and key pieces of information related to using those individual resources with same, could at least enable more sophisticated use of those resources if not increased discovery.

    (You’ve got your ERM in my discovery interface! No, you got your discovery interface in my ERM! Er… guess that doesn’t quite translate.)

  10. Flickr Mosaic: Phyllotaxy (cc:by-nc-sa); Librarians-Haunted-Love (cc:by-nc-sa); Square Peg (cc:by-nc-sa); The Burden of Thought (cc:by-nc)

CIL 2009: Social Network Profile Management

Speaker: Greg Schwartz

Who are you online? Identity is what you say about you and what others say about you. However, it’s more than just that. It includes the things you buy, the tools you use, the places you spend your time, etc.

You do not own your online identity. You can’t control what people find out about you, but you can influence it.

  1. Own your user name. Pick one and stick to it. Even better if you can use your real name. (checkusernames.com)
  2. Join the conversation. Develop your identity by participating in social networks.
  3. Listen. Pay attention to what other people are saying about you.
  4. Be authentic. Ultimately, social networking is about connecting your online identity to your in-person identity.

Speaker: Michael Porter

MP was the project manager for the social tools on WebJunction. It’s designed to be for librarians and library staff.

If you are representing your organization online, be yourself, but also be sensitive to how that could be perceived. Share your library success stories!

Speaker: Sarah Houghton-Jan

Library online identities should be created with a generic email address, should be up-to-date, and should allow comment and interaction with users. Keep the tone personable.

Don’t use multiple identities. Make sure that someone is checking the contact points. You’ll get better results if you disperse the responsibility for library online identities across your institution rather than relying on one person to manage it all.

Speaker: Amanda Clay Powers

People have been telling their stories for a long time, and online social networks are just another tool for doing that. Some people are more comfortable with this than others. It’s our role to educate people about how to manage their online identities, however, our users don’t always know that librarians can help them.

On Facebook, you can manage social data by creating friends lists. This functionality is becoming more important as online social networks grow and expand.

CIL 2009: New Strategies for Digital Natives

Speaker: Helene Blowers

Begins with a video of a 1yo. unlocking and starting up a Preschool Adventure game on an iPhone, and then paging through images in the photo gallery. Joey is a digital native and the future of library users.

Digital natives are those born after 1980. When they were 1, IBM distributed the first commercial PC. Cellular phones were introduced at the age of 3. By the time they were 14, the internet was born.

Web 1.0 was built on finding stuff, Web 2.0 was built on connecting with other users and share information. Digital natives are used to not only having access to the world through the internet, but also engaging with it.

Business Week categorized users by how they interact with the internet and their generation. This clearly lays out the differences between how the generations use this tool, and it should inform the way we approach library services to them.

Digital native realities:

  • Their identity online is the same as their in-person identity. They grew up with developing both at the same time, as oppose to those who came before. Facebook, MySpace, Twitter, Flixster, and LinkedIn are the top five online social networks, according to a report in January. How many of them do you have an identity in?
  • The ability to create and leave your imprint somewhere is important to digital natives. According to the Pew Internet & American Life, those who participate in social networks are more likely to create unique content than those who do not.
  • We are seeing a shift from controlled information to collaborative information, so digital information quality has become important and a personal responsibility to digital natives. After a study showed that Wikipedia was as accurate as Britannica resulted in EB adding a wiki layer to their online presence.
  • Digital natives have grown up in a world they believe to be safe, whether it is or not. Less than 0.08% of students say that they have met someone online without their parents knowledge, and about 65% say that they ignored or deleted strangers that tried to contact them online. However, that doesn’t stop them from intentionally crossing that line in order to rebel against rules.
  • Digital opportunity is huge. There are no barriers, the playing field has been leveled, access is universal, connection ubiquitous, and it’s all about me.
  • Digital sharing is okay. It’s just sharing. They aren’t concerned with copyright or ownership. Fanfic, mashups, remixes, parodies… Creative Commons has changed the way we look at ownership and copyright online.
  • Privacy online and in their social networks is not much of a concern. Life streams aggregate content from several social networks, providing the big picture of someone’s online life.
  • What you do online makes a difference — digital advocacy. This was clear during the US presidential election last year.

What does this mean for libraries? How do we use this to support the information needs of our users?

Think about ways to engage with virtual users — what strategies do we need in order to connect library staff and services with users in meaningful ways? Think about ways to enrich the online experience of users that then enhances their experiences in the physical library and their daily lives. Think about ways to empower customers to personalize and add value to their library experience so that they feel good about themselves and their community.

css.php