Speaker: Emily Guhde, NCLIVE
“We’ve Got Your Number: Making Usage Data Matter” is the project they are working on. What is a good target cost per use for their member libraries? They are organizing this by peer groups. How can the member libraries improve usage? They are hoping that other libraries will be able to replicated this in the future.
Speaker: Francis Kayiwa, UIC
He is a server administrator with library training, and wanted to be here to understand what it is his folks are coming back and asking him to do. Cross-pollinate conferences — try to integrate other kinds of conferences happening nearby.
Speaker: Annette Bailey, Virginia Tech
Co-developed LibX with her husband, now working on a new project to visualize what users are clicking on after they get a search result in Summon. This is a live, real-time visualization, pulled from the Summon API.
Speaker: Angie Rathnel, University of Kansas
Have been using a SAS called Callisto to track and claim eresources. It tracks access to entitlements daily/weekly, and can check to make sure proxy configurations are set up correctly.
Speaker: Cindy Boeke, Southern Methodist University
Why aren’t digital library collections included with other library eresources on lists and such (like the ubiquitous databases A-Z page)?
Speaker: Rick Burke, SCELC
SIPX to manage copyright in a consortial environment. Something something users buying access to stuff we already own. I’m guessing this is more for off-campus access?
Speaker: Margy Avery, MIT Press
Thinking about rich/enhanced digital publications. Want to work with libraries to make this happen, and preservation is a big issue. How do we catalog/classify this kind of resource?
Speaker: Jason Price, Claremont Colleges
Disgruntled with OpenURL and the dependency on our KB for article-level access. It is challenging to keep our lists (KBs) updated and accurate — there has to be a better way. We need to be working with the disgrundterati who are creating startups to address this problem. Pubget was one of the first, and since then there is Dublin Six, Readcube, SIPX, and Callisto. If you get excited about these things, contact the startups and tell them.
Speaker: Wilhelmina Ranke, St. Mary’s University
Collecting mostly born digital collections, or at least collections that are digitized already, in the repository: student newspaper, video projects, and items digitized for classroom use that have no copyright restrictions. Doesn’t save time on indexing, but it does save time on digitizing.
Speaker: Bonnie Tijerina, Harvard
The #ideadrop house was created to be a space for librar* to come together to talk about librar* stuff. They had a little free library box for physical books, and also a collection of wireless boxes with free digital content anyone could download. They streamed conversations from the living room 5-7 times a day.
Speaker: Rachel Frick Digital Public Library of America focuses on content that is free to all to create a more informed citizenry. They want to go beyond just being a portal for content. They want to be a platform for community involvement and conversations.
Speakers: Dani Roach & Carolyn DeLuca, University of Saint Thomas
We’re all very familiar with print serials cancellation projects, but now we’re starting to see this in the electronic world, particularly as more libraries are walking away from big deals. There are lots of documents talking about the life-cycle of an eresource, but cancellation hasn’t really been addressed until recently, with the TERMS section covering it along with others.
As much work as it is to get into a relationship with a vendor/resource, it’s just has hard to leave it.
Eresource breakups are often caused by the renewal time, low use, high cost, etc. The heart-breaker could be internal players like the liaison or ERMS staff acting as the divorce lawyer once the decision was made, or external players like the publisher/provider/broker/consortium/vendor.
The three types of breakups are cancelled, ceased, or migrated. In each case, you need to assess the status, the holdings, the platform, and the provider. There are some things we have control over, but there are many more things outside of our control, due to the marketplace around information. Don’t take it personally, libraries. It’s not about you.
There are tools where you make the change (catalog, ERM, course management system, and paper files, proxy config, local tools, customized holdings, archives/Portico/local), and then there are tools where you announce the change (LibGuides, RefWorks, site-wide search metadata, blogs & other social media).
UKSG created a cancellation form, which looks lovely. And sadly, our ERMS can’t track everything.
Vendors can play the role of best friend. They want to help, and make sure we’re making the right decision so we don’t come back crying on their shoulder later. Some tips: Know who signed the license, confirm post-cancellation rights, do not count on a refund, plan cancellations and migrations well in advance of renewal, and know what you want (if replacing a product, do your research).
Considerations from the publisher/provider: ownership transfers, licensing, grace periods, overlap periods, and personal relationships.
When does it end? In general, the library gets to decide when to shut it down. It could be immediate, or you could wait until the end of the semester or the end of the subscription term. Sometimes resources are put on probation and given some time to demonstrate value to the community.
What’s left? Post-cancellation access? Deleting or weeding — did you ask the provider to remove the old edition? When a product has migrated, you’ll need to change the tutorials, screenshots, videos, etc.
What the “Google Generation” Says About Using Library & Information Collections, Services, and Systems in the Digital Age
Speaker: Michael Eisenberg, University of Washington Information School
We’ve moved from scarcity to abundance to overload. We have so many great resources our students don’t know where to begin. They’re overwhelmed.
Think about how our computing technology has evolved and shrank in both size and price while increasing in power over the past 30 years. Where will be 20 years from now?
We live in a parallel information universe that is constantly feeding information back to us. The library is anywhere anytime, so how can we best meet the information needs of our users?
Project Information Literacy seeks to answer what it means to be a student in the digital age. They have been assessing different types of students on how they find and use information to get generalized pictures of who they are.
Why, when you have an information need, do you turn to Google first and not research databases?
Students ignore faculty warnings about Wikipedia. They still use it, but they just don’t cite it.
Students aren’t really procrastinators, they’re just busy. They are working to the last minute because every minute is highly scheduled. Have we changed our staffing or the nature of our services to help them at point of need?
Students don’t think of librarians as people who can help them with their research, they think of them as people who can help them with resources. They are more likely to go to their instructors and classmates before librarians during the research process. The hardest part for them is getting started and defining the topic (and narrowing it down). They don’t think librarians can help them with that, even though we can, and do (or should if we aren’t already).
Students are more practiced at writing techniques than research strategies. Professors complain that students can’t write, but maybe writing shouldn’t be the only method of expression.
Most students don’t fully understand the research process and what is expected. They need clarity on the nature and scope of assignments, and they aren’t used to critical thinking (“just tell me what you want and I’ll give it to you”). Most handouts from profs don’t explain this well, focusing more on mechanics and sending students to the library shelves (and not to databases or online resources). Rarely do they suggest talking to the librarian.
Students are not the multi-taskers we think they are, particularly during crunch time. Often they will use the library and library computers to force themselves to limit the distractions and focus. They use Facebook breaks as incentives to get things done.
After they graduate, former students are good with technology, but not so good with low-tech, traditional research/information discovery skills.
Information literacy needs are more important than ever, but they are evolving. Search to task to use to synthesis to evaluation — students need to be good at every stage. The library is shifting from the role of information to space, place, and equipment. Buying the resources is less of an emphasis (although not less in importance), and the needs change with the academic calendar.
What do we do about all this?
Infuse high quality, credible resources and materials into courses and classes. Consider resources and collections in relation to Wikipedia. Infuse information literacy learning opportunities into resources, access systems, facilities, and services (call it “giving credit,” which they understand more than citing). Provide resources, expertise, and services related to assignments. Re-purpose staff and facilities related to calendar and needs. Offer to work with faculty to revise handouts — emphasize the quality of resources not the mechanics. Offer flexible and collaborative spaces with a range of capabilities and technology, less emphasis on print collection development. Consider school-to-work transitions in access systems, resources, services, and instruction.
Beyond formal instruction, what are the ways we can help students gain the essential information literacy skills they need? That is the challenge for eresources librarians.
Conference season, or at least the part of it that appeals to my area of librarianship, is starting soon.
Up first for me is Computers in Libraries in DC, where I won’t be attending, but instead vacationing nearby (since it is so close) and visiting with colleagues and friends who will be attending. I’d go, but I already have funding this year for three conferences, and it didn’t seem fair to ask for another.
Next, I fly to Austin for the Electronic Resources & Libraries conference. From the venue to the content, this is becoming my favorite conference. I’ve had to actively introduce more diversity to the sessions I choose to attend, otherwise I would spend the whole conference geeking out about use data and spreadsheets and such.
Finally, I head to Buffalo for the conference that shaped me into the librarian I became: NASIG (North American Serials Interest Group). I like this one because I’ve known many of the attendees for the entirety of my relatively short career, and because it works very hard to not be just a librarian conference, but rather an industry-wide discussion of all things serial in libraryland.
It was in the context of thinking about these upcoming conferences that I read the latest Prof Hacker blog post from The Chronicle of Higher Education. Ryan Cordell writes about his experiences with conference tweeting and the recent revelations he has had regarding the impact this can have on the presenters, whether they are active participants on Twitter or not. Many things he wrote resonated with me, and reminded me that Twitter — as well as other popular social media platforms — is no longer the private back-channel of a few techie friends, but is a global platform that can have a broader impact than any of us may know.
I suggest reading the whole article, but I would like to quote here the Principles of Conference Tweeting that Cordell offers, as something for us all to keep in mind:
I will post praise generously, sharing what I find interesting about presentations.
Likewise, I will share pertinent links to people and projects, in order to bring attention to my colleagues’ work.
When posting questions or critiques, I will include the panelist’s username (an @ mention) whenever possible.
If the panelist does not have a username—or if I cannot find it—I will do my best to alert them when I post questions or critiques, rather than leaving them to discover those engagements independently.
I will not post questions to Twitter that I would not ask in the panel Q&A.
I will not use a tone on Twitter that I would not use when speaking to the scholar in person.
I will avoid “crosstalk”—joking exchanges only tangentially related to the talk—unless the presenter is explicitly involved in the chatter.
I will refuse to post or engage with posts that comment on the presenter’s person, rather than the presenter’s ideas.
Speaker: Susan Stearns, VP of Strategic Partnerships of Ex Libris Group
Both library as a percentage of university expenditures and the number of library staff per student have been going down. The percentage of library expenditures spent on electronic resources has been going up dramatically.
There is a need to eliminate the duplication of data and workflows, and the silo systems in libraries today. Alma intends to unify both the data and the data environment: acquisitions, metadata management, fulfillment, and analytics.
Collaborative metadata management is a hybrid model to balance global sharing with local needs. In English, this means you can have a catalog that includes both an inventory of locally owned items and a collection of items shared by one or more “communities.” Multiple metadata schema are supported within the system in their native formats — no crosswalks required.
Individual library staff users can set up “home pages” within the system that includes widgets with data, alerts, and reports. This can help with making decisions about the collection. Analytics are also embedded directly in the workflow (i.e. a graph representing the balance remaining in a fund displayed when an order using that fund is viewed/entered).
Speaker: Maria Bunevski, Ex Libris
Preparation for moving to a new system, particularly a radically new system like Alma, requires spending some time thinking about workflows, data, technical aspects (integration points, etc.), and training.
Project initiation phase requires a lot of training sessions to fully grasp all of the change that needs to happen.
The implementation phase involves a mix of on-site work and remote tweaking. At some point work has to freeze in the old system before cutting over to the new one.
VCU is currently in the post-implementation phase. This is the point where un-configured things are discovered, along with gaps in workflow.
Speaker: John Duke, VCU Libraries
They had Aleph, SFX, Verde, MetaLib, Primo, ARC, ILLiad, university systems, etc. before, and they wanted to bring the functions together. They didn’t end up with a monolithic system for everything, but they got closer.
Workflows and other aspects have been simplified.
The system is not complete, either because Ex Libris hadn’t thought of it or because VCU hasn’t figured out how to incorporate it. Internet outages, security issues, and conceptual difficulties have thrown up road blocks along the way.
“Educational Utility Computing: Perspectives on .edu and the Cloud”
Mark Ryland, Chief Solutions Architect at Amazon Web Services
AWS has been a part of revolutionizing the start-up industries (i.e. Instagram, Pinterest) because they don’t have the cost of building server infrastructures in-house. Cloud computing in the AWS sense is utility computing — pay for what you use, easy to scale up and down, and local control of how your products work. In the traditional world, you have to pay for the capacity to meet your peak demand, but in the cloud computing world, you can level up and down based on what is needed at that moment.
Economies, efficiencies of scale in many ways. Some obvious: storage, computing, and networking equipment supply change; internet connectivity and electric power; and data center sitting, redundancy, etc. Less obvious: security and compliance best practices; datacenter internal innovations in networking, power, etc.
AWS and .EDU: EdX, Coursera, Texas Digital Library, Berkeley AMP Lab, Harvard Medical, University of Phoenix, and an increasing number of university/school public-facing websites.
Expects that we are heading toward cloud computing utilities to function much like the electric grid — just plug in and use it.
“Libraries in Transition”
Marshall Breeding, library systems expert
We’ve already seen the shift of print to electronic in academic journals, and we’re heading that way with books. Our users are changing in the way they expect interactions with libraries to be, and the library as space is evolving to meet that, along with library systems.
Web-based computing is better than client/server computing. We expect social computing to be integrated into the core infrastructure of a service, rather than add-ons and afterthoughts. Systems need to be flexible for all kinds of devices, not just particular types of desktops. Metadata needs to evolve from record-by-record creation to bulk management wherever possible. MARC is going to die, and die soon.
How are we going to help our researchers manage data? We need the infrastructure to help us with that as well. Semantic web — what systems will support it?
Cooperation and consolidation of library consortia; state-wide implementations of SaaS library systems. Our current legacy ILS are holding libraries back from being able to move forward and provide the services our users want and need.
A true cloud computing system comes with web-based interfaces, externally hosted, subscription OR utility pricing, highly abstracted computing model, provisioned on demand, scaled according to variable needs, elastic.
“Moving Up to the Cloud”
Mark Triest, President of Ex Libris North America
Currently, libraries are working with several different systems (ILS, ERMS, DRs, etc.), duplicating data and workflows, and not always very accurately or efficiently, but it was the only solution for handling different kinds of data and needs. Ex Libris started in 2007 to change this, beginning with conversations with librarians. Their solution is a single system with unified data and workflows.
They are working to lower the total cost of ownership by reducing IT needs, minimize administration time, and add new services to increase productivity. Right now there are 120+ institutions world-wide who are in the process of or have gone live with Alma.
Automated workflows allow staff to focus on the exceptions and reduce the steps involved.
Descriptive analytics are built into the system, with plans for predictive analytics to be incorporated in the future.
Future: collaborative collection development tools, like joint licensing and consortial ebook programs; infrastructure for ad-hoc collaboration
“Cloud Computing and Academic Libraries: Promise and Risk”
John Ulmschneider, Dean of Libraries at VCU
When they first looked at Alma, they had two motivations and two concerns. They were not planning or thinking about it until they were approached to join the early adopters. All academic libraries today are seeking to discover and exploit new efficiencies. The growth of cloud-resident systems and data requires academic libraries to reinvigorate their focus on core mission. Cloud-resident systems are creating massive change throughout out institutions. Managing and exploiting pervasive change is a serious challenge. Also, we need to deal with security and durability of data.
Cloud solutions shift resources from supporting infrastructure to supporting innovation.
Efficiencies are not just nice things, they are absolutely necessary for academic libraries. We are obligated to upend long-held practice, if in doing so we gain assets for practice essential to our mission. We must focus recovered assets on the core library mission.
Agility is the new stability.
Libraries must push technology forward in areas that advance their core mission. Infuse technology evolution for libraries with the values needs of libraries. Libraries must invest assets as developers, development partners, and early adopters. Insist on discovery and management tools that are agnostic regarding data sources.
Managing the change process is daunting.. but we’re already well down the road. It’s not entirely new, but it does involve a change in culture to create a pervasive institutional agility for all staff.
Updates from Serials Solutions – mostly Resource Manager (Ashley Bass):
Keep up to date with ongoing enhancements for management tools (quarterly releases) by following answer #422 in the Support Center, and via training/overview webinars.
Populating and maintaining the ERM can be challenging, so they focused a lot of work this year on that process: license template library, license upload tool, data population service, SUSHI, offline date and status editor enhancements (new data elements for sort & filter, new logic, new selection elements, notes), and expanded and additional fields.
Workflow, communication, and decision support enhancements: in context help linking, contact tool filters, navigation, new Counter reports, more information about vendors, Counter summary page, etc. Her most favorite new feature is “deep linking” functionality (aka persistent links to records in SerSol). [I didn’t realize that wasn’t there before — been doing this for my own purposes for a while.]
Next up (in two weeks, 4th quarter release): new alerts, resource renewals feature (reports! and checklist!, will inherit from Admin data), Client Center navigation improvements (i.e. keyword searching for databases, system performance optimization), new license fields (images, public performance rights, training materials rights) & a few more, Counter updates, SUSHI updates (making customizations to deal with vendors who aren’t strictly following the standard), gathering stats for Springer (YTD won’t be available after Nov 30 — up to Sept avail now), and online DRS form enhancements.
In the future: license API (could allow libraries to create a different user interface), contact tools improvements, interoperability documentation, new BI tools and reporting functionality, and improving the Client Center.
Also, building a new KB (2014 release) and a web-scale management solution (Intota, also coming 2014). They are looking to have more internal efficiencies by rebuilding the KB, and it will include information from Ulrich’s, new content types metadata (e.g. A/V), metadata standardization, industry data, etc.
Summon Updates (Andrew Nagy):
I know very little about Summon functionality, so just listened to this one and didn’t take notes. Take-away: if you haven’t looked at Summon in a while, it would be worth giving it another go.
360 Link Customization via JavaScript and CSS (Liz Jacobson & Terry Brady, Georgetown University):
Goal #1: Allow users to easily link to full-text resources. Solution: Go beyond the out-of-the box 360 Link display.
Goal #2: Allow users to report problems or contact library staff at the point of failure. Solution: eresources problem report form
They created the eresources problem report form using Drupal. The fields include contact information, description of the resource, description of the problem, and the ability to attach a screenshot.
When they evaluated the slightly customized out of the box 360 Link page, they determined that it was confusing to users, with too many options and confusing links. So, they took some inspiration from other libraries (Matthew Reidsma’s GVUS jQuery code available on Github) and developed a prototype that uses custom JavaScript and CSS to walk the user through the process.
Some enhancements included: making the links for full-text (article & journal) butttons, hiding additional help information and giving some hover-over information, parsing the citation into the problem report page, and moving the citation below the links to full-text. For journal citations with no full-text, they made the links to the catalog search large buttons with more text detail in them.
Some of the challenges of implementing these changes is the lack of a test environment because of the limited preview capablities in 360 Link. Any changes actually made required an overnight refresh and they would be live, opening the risk of 24 hour windows of broken resource links. So, they created their own test environment by modifying test scenarios into static HTML files and wrapping them in their own custom PHP to mimic the live pages without having to work with the live pages.
[At this point, it got really techy and lost me. Contact the presenters for details if you’re interested. They’re looking to go live with this as soon as they figure out a low-use time that will have minimal impact on their users.]
Customizing 360 Link menu with jQuery (Laura Wrubel, George Washington University)
They wanted to give better visual clues for users, emphasize the full-text, have more local control over linkns, and visual integration with other library tools so it’s more seamless for users.
They started with Reidsma’s code, then then forked off from it. They added a problem link to a Google form, fixed ebook chapter links and citation formatting, created conditional links to the catalog, and linked to their other library’s link resolver.
They hope to continue to tweak the language on the page, particularly for ILL suggestion. The coverage date is currently hidden behind the details link, which is fine most of the time, but sometimes that needs to be displayed. They also plan to load the print holdings coverage dates to eliminate confusion about what the library actually has.
In the future, they would rather use the API and blend the link resolver functionality with catalog tools.
Custom document delivery services using 360 Link API (Kathy Kilduff, WRLC)
They facilitate inter-consortial loans (Consortium Loan Service), and originally requests were only done through the catalog. When they started using SFX, they added a link there, too. Now that they have 360 Link, they still have a link there, but now the request form is prepopulated with all of the citation information. In the background, they are using the API to gather the citation information, as well as checking to see if there are terms of use, and then checking to see if there are ILL permissions listed. They provide a link to the full-text in the staff client developed for the CLS if the terms of use allow for ILL of the electronic copy. If there isn’t a copy available in WRLC, they forward the citation information to the user’s library’s ILL form.
License information for course reserves for faculty (Shanyun Zhang, Catholic University)
Included course reserve in the license information, but then it became an issue to convey that information to the faculty who were used to negotiating it with publishers directly. Most faculty prefer to use Blackboard for course readings, and handle it themselves. But, they need to figure out how to incorporate the library in the workflow. Looking for suggestions from the group.
Advanced Usage Tracking in Summon with Google Anaytics (Kun Lin, Catholic University)
In order to tweak user experience, you need to know who, what, when, how, and most important, what were they thinking. Google Anayltics can help figure those things out in Summon. Parameters are easy ways to track facets, and you can use the data from Google Analytics to figure out the story based on that. Tracking things the “hard way,” you can use the conversion/goal function of Google Analytics. But, you’ll need to know a little about coding to make it work, because you have to add some javascripts to your Summon pages.
Use of ERM/KB for collection analysis (Mitzi Cole, NASA Goddard Library)
Used the overlap analysis to compare print holdings with electronic and downloaded the report. The partial overlap can actually be a full overlap if the coverage dates aren’t formatted the same, but otherwise it’s a decent report. She incorporated license data from Resource Manager and print collection usage pulled from her ILS. This allowed her to create a decision tool (spreadsheet), and denoted the print usage in 5 year increments, eliminating previous 5 years use with each increment (this showed a drop in use over time for titles of concern).
Discussion of KnowledgeWorks Management/Metadata (Ben Johnson, Lead Metadata Librarian, SerialsSolutions)
After they get the data from the provider or it is made available to them, they have a system to automatically process the data so it fits their specifications, and then it is integrated into the KB.
They deal with a lot of bad data. 90% of databases change every month. Publishers have their own editorial policies that display the data in certain ways (e.g., title lists) and deliver inconsistent, and often erroneous, metadata. The KB team tries to catch everything, but some things still slip through. Throught the data ingestion process, they apply rules based on past experience with the data source. After that, the data is normalized so that various title/ISSN/ISBN combinations can be associated with the authority record. Finally, the data is incorporated into the KB.
Authority rules are used to correct errors and inconsistencies. Rule automatically and consistently correct holdings, and they are often used to correct vendor reporting problems. Rules are condified for provider and database, with 76,000+ applied to thousands of databases, and 200+ new rules are added each month.
Why does it take two months for KB data to be corrected when I report it? Usually it’s because they are working with the data providers, and some respond more quickly than others. They are hoping that being involved with various initiatives like KBART will help fix data from the provider so they don’t have to worry about correcting it for us, but also making it easier to make those corrections by using standards.
Client Center ISSN/ISBN doesn’t always work in 360 Links, which may have something to do with the authority record, but it’s unclear. It’s possible that there are some data in the Client Center that haven’t been normalized, and could cause this disconnect. And sometimes the provider doesn’t send both print and electronic ISSN/ISBN.
What is the source for authority records for ISSN/ISBN? LC, Bowker, ISSN.org, but he’s not clear. Clarification: Which field in the MARC record is the source for the ISBN? It could be the source of the normalization problem, according to the questioner. Johnson isn’t clear on where it comes from.
Speakers: Matt Torrence, Audrey Powers, & Megan Sheffield, University of South Florida
Are collection development policies viable today? In order answer this, they sent out a survey to ARL libraries to see if they are using them or if they’re experimenting with something else. They were also interested to know when and how data is being used in the process.
The survey results will be published in the proceedings. I will note anything here that seems particularly interesting, but it looks like all they are doing now is reading that to us.
Are collection development policies being used? Yes, sort of. Although most libraries in the survey do have them, they tend to be used for accreditation and communication, and often they are not consistently available either publicly or internally.
What are the motivations for using collection development policies? Tends to be more for external/marketing than for internal workflows.
They think that a collection development “philosophy” may be a more holistic response to the changing nature of collection development.
Speakers: two people from the University of Arkansas at Little Rock, but they had four names on the PPT, and I didn’t catch who was who
They recently decided to revise their collection development policy/guidelines based on a recommendation from a strategic planning ARL Collection Analysis Project. They also had quite a few new librarians who needed to work with faculty selectors.
They did a literature review and gathered information on practices from peer institutions. They actually talked to the Office of Institutional Research about data on academic degree programs. And, like students, they looked online to see if they could borrow from existing documents.
One thing they took away from the review of what other libraries have out there was that they needed to have the document live on the web, and not just on paper in a binder in someone’s office.
Policies/guidelines should be continuously updating, flexible, acknowledge consortia memberships, acknowledge new formats, and strike a balance between being overly detailed and too general.
They see that the project has had some benefits, not only to themselves but also to provide a guide for current and future users of the policies. It is also a valuable tool for transmitting institutional memory.
And so does the WordPress app for iPad, or at least the current version. I had drafts of the three sessions I attended this afternoon, ready to publish as soon as I returned to my room, which is the only place I can connect to the wifi. As soon as the WordPress connected to update, the contents of all three posts reverted to the blank drafts I had created as placeholders.
Yeah. Pissed. That’d be me right now.
In short:
Eresources librarians need to demonstrate their value to the library/university, and they either need more staff to do the increasing work, or other departments need to suck it up and process e-stuff like they should. And yes, someone needs to handle licensing, but that someone shouldn’t also be responsible for every little tiny detail of eresources management (i.e. cataloging, trouble-shooting, invoices, etc.) when there are staff already handling similar processes for other materials.
Librarians need to learn how to market eresources effectively, and assess their marketing strategies effectively. Marie Kennedy has a book coming out next year that can help you with that.
Eresources librarians (or licensing librarians) need to make sure language supporting text mining is included in their license agreements with publishers. Your researchers will thank you for it later, and your future self will be happy to not have to go back and renegotiate it into existing contracts.
A couple weeks ago I gave a presentation on our ERMS implementation at the annual meeting of the Potomac Technical & Processing Librarians. It was a great honor to be asked to present along with a panel of other librarians talking about their ERMS. It was also the longest I’ve been asked to speak, and while a completely nerve-wracking experience internally, it seemed to have been well-received. If you are interested, my notes and slides are now available on SlideShare.