CiL 2008: Speed Searching

Speaker: Greg Notess

His talk summarizes points from his Computers in Libraries articles on the same topic, so go find them if you want more details than what I provide.

It takes time to find the right query/database, and to determine the best terminology to use in order to find what you are seeking. Keystroke economy makes searching faster, like the old OCLC FirstSearch 3-2-2-1 searching. Web searching relevancy is optimized by using only a few unique words rather than long queries. Do spell checking through a web search and then take that back into a reference database. Search suggestions on major search engines help with the spelling problem, and the suggestions are ranked based on the frequency with which they are searched, but they require you to type slowly to use them effectively and increase your search speed. Copy and paste can be enhanced through browser plugins or bookmarklets that allow for searching based on selected text.

The search terms matter. Depending on the source, average query length searches using unique terms perform better over common search terms or long queries. Use multiple databases because it’s fun, you’re a librarian, and there is a lack of overlap between data sources.

Search switching is not good for quick look-ups, but it can be helpful with hard to find answers that require in-depth query. We have a sense that federated searching should be able to do this, but some resources are better searched in their native interfaces in order to find relevant sources. There are several sites that make it easy to switch between web search engines using the same query, including a nifty site that will allow you to easily switch between the various satellite mapping sources for any location you choose.

I must install the Customize Google Firefox plugin. (It’s also available for IE7, but why would you want to use IE7, anyway?)

CiL 2008 Keynote: Libraries as Happiness Engines

Speaker: Elizabeth Lane Lawley, Rochester Institute of Technology

Libraries are an emotional center of a community that make people happy. The elements of happiness include: satisfying work to do, the experience of being good at something, time spent with people we like, and a chance to be a part of something bigger. Libraries will survive if we remain in the happiness business.

Virtuality is a way of beating an unhappy life. People go into virtual world to escape a lack of happiness in their real life. And, there is a blurring of the boundaries between virtual worlds and real worlds as players make connections in both that bleed over into the other.

“The grind” is a process that gamers go through in order to move forward or advance in levels. Players do tasks in game repeatedly until their goals are achieved. Lawley thinks of the grind as being a meditative activity, due in part to the repetitive nature of it. Why can’t we convey this to our potential readers? We can correlate the reward of working through a game to the reward of reading a book through to the end.

Tupperware parties are an example of real world games, with competition and rewards. Salespeople will work hard to move up levels and be recognized for their work. Another real world game that is popular in public libraries is Super Sleuth, which gives daily clues that kids research to find the answer every week. Summer reading programs are also like games, and they make reading a challenge that kids want to do in order to reach a point goal to win a prize.

Chore Wars is an online game that blurs the line between real and virtual worlds. Parents post job tasks with point values, and kids earn those points by doing the chores. In the business world, Seriosity’s Attent is a game that causes players to evaluate the value of email message sent and received. Social Genius takes casual game concepts and applies them to enterprise problems/solutions, basically by making players learn the names and faces of other people in their organization and forcing them to keep their online photos and bios up-to-date.

Passively Multiplayer Online Gaming gives points to players for visiting websites, among other things. Missions become annotated pathfinders that reward players for surfing websites that have some sort of relevancy to each other. It makes the process of going to websites fun instead of tedious.

Games can serve as gateway drugs, like Guitar Hero. [side note: I heard recently about someone modifying a real guitar to use the strum bar and buttons used on the Guitar Hero guitar.]

Online rebound is what happens when we go from virtual to real and back again, like LAN parties, Moo cards (“We love the web, but you can’t put it in your pocket.”), and Etsy (hand-crafted items sold online). Virtual does not take the tangible away. You want to retain a connection to the real world. Libraries are taking advantage of this online rebound by creating spaces where people can be together physically while also being online.

How does your library make people feel happy? How does it pull them into something bigger than themselves that makes them feel playful and productive at the same time?

CiL 2008: Catalog Effectiveness

Speaker: Rebekah Kilzer

The Ohio State University Libraries have used Google Analytics for assessing the use of the OPAC. It’s free for sites up to five million page views per month — OSU has 1-2 million page views per month. Libraries would want to use this because most integrated library systems offer little in the way of use statistics, and what they do have isn’t very… useful. You will need to add some code that will display on all OPAC pages.

Getting details about how users interact with your catalog can help with making decisions about enhancements. For example, knowing how many dial-up users interact with the site could determine whether or not you want to develop style sheets specifically for them, for example. You can also track what links are being followed, which can contribute to discussions on link real estate.

There are several libraries that are mashing up Google Analytics information with other Google tools.


Speakers: Cathy Weng and Jia Mi

The OPAC is a data-centered, card-catalog retrieval system that is good for finding known items, but not so good as an information discovery tool. It’s designed for librarians, not users. Librarian’s perceptions of users (forgetful, impatient) prevents librarians from recognizing changes in user behavior and ineffective OPAC design.

In order to see how current academic libraries represent and use OPAC systems, they studied 123 ARL libraries’ public interfaces and search capabilities as well as their bibliographic displays. In the search options, two-thirds of libraries use “keyword” as the default and the other third use “title.” The study also looked at whether or not the keyword search was a true keyword search with an automatic “and” or if the search was treated as a phrase. Few libraries used relevancy ranking as the default search results sorting.

There are some great disparities in OPAC quality. Search terms and search boxes are not retained on the results page, post-search limit functions are not always readily available, item status are not available on search results page, and the search keywords are not highlighted. These are things that the most popular non-library search engines do, which is what our users expect the library OPAC to do.

Display labels are MARC mapping, not intuitive. Some labels are suitable for certain types of materials but not all (proper name labels for items that are “authored” by conferences). They are potentially confusing (LCSH & MeSH) and occasionally inaccurate. The study found that there were varying levels of effort put to making the labels more user-friendly and not full of library jargon.

In addition to label displays, OPACs also suffer from the way the records are displayed. The order of bibliographic elements effect how users find relevant information to determine whether or not the item found is what they need.

There are three factors that contribute to the problem of the OPAC: system limitations, libraries not exploiting full functionality of ILS, and MARC standards are not well suited to online bibliographic display. We want a system that doesn’t need to be taught, that trusts users as co-developers, and we want to maximize and creatively utilize the system’s functionality.

The presentation gave great examples of why the OPAC sucks, but few concrete examples of solutions beyond the lipstick-on-a-pig catalog overlay products. I would have liked to have a list of suggestions for label names, record display, etc., since we were given examples of what doesn’t work or is confusing.

CiL 2008: Woepac to Wowpac

Moderator: Karen G. Schneider – “You’re going to go un-suck your OPACs, right?”


Speaker: Roy Tennant

Tennant spent the last ten years trying to kill off the term OPAC.

The ILS is your back end system, which is different from the discovery system (doesn’t replace the ILS). Both of these systems can be locally configured or hosted elsewhere. Worldcat Local is a particular kind of discovery system that Tenant will talk about if he has time.

Traditionally, users would search the ILS to locate items, but now the discovery system will search the ILS and other sources and present it to the user in a less “card catalog” way. Things to consider: Do you want to replace your ILS or just your public interface? Can you consider open source options (Koha, Evergreen, vuFind, LibraryFind etc.)? Do you have the technical expertise to set it up and maintain it? Are you willing to regularly harvest data from your catalog to power a separate user interface?


Speaker: Kate Sheehan

Speaking from her experience of being at the first library to implement LibraryThing for Libraries.

The OPAC sucks, so we look for something else, like LibraryThing. The users of LibraryThing want to be catalogers, which Sheehan finds amusing (and so did the audience) because so few librarians want to be catalogers. “It’s a bunch of really excited curators.”

LibraryThing for libraries takes the information available in LibraryThing (images, tags, etc.) and drops them into the OPAC (platform independent). The display includes other editions of books owned by the library, recommendations based on what people actually read, and a tag cloud. The tag cloud links to a tag browser that opens up on top of the catalog and allows users to explore other resources in the catalog based on natural language tags rather than just subject headings. Using a Greasmonkey script in your browser, you can also incorporate user reviews pulled from LibraryThing. Statistics show that the library is averaging around 30 tag clicks and 18 recommendations per day, which is pretty good for a library that size.

“Arson is fantastic. It keeps your libraries fresh.” — Sheehan joking about an unusual form of collection weeding (Danbury was burnt to the ground a few years ago)

Data doesn’t grow on trees. Getting a bunch of useful information dropped into the catalog saves staff time and energy. LibraryThing for Libraries didn’t ask for a lot from patrons, and it gave them a lot in return.


Speaker: Cindi Trainor

Are we there yet? No. We can buy products or use open source programs, but they still are not the solution.

Today’s websites are consist of content, community (interaction with other users), interactivity (single user customization), and interoperability (mashups). RSS feeds are the intersection of interactivity and content. There are a few websites that are in the sweet spot in the middle of all of these: Amazon (26/32)*, Flickr (26/32), Pandora (20/32), and Wikipedia (21/32) are a few examples.

Where are the next generation catalog enhancements? Each product has a varying degree of each element. Using a scoring system with 8 points for each of the four elements, these products were ranked: Encore (10/32), LibraryFind (12/32), Scriblio (14/32), and WorldCat Local (16/32). Trainor looked at whether the content lived in the system or elsewhere and the degree to which it pulled information from sources not in the catalog. Library products still have a long way to go – Voyager scored a 2/32.

*Trainor’s scoring system as described in paragraph three.


Speaker: John Blyberg

When we talk about OPACs, we tend to fetishize them. In theory, it’s not hard to create a Wowpac. The difficulty is in creating the system that lives behind it. We have lost touch with the ability to empower ourselves to fix the problems we have with integrated library systems and our online public access catalogs.

The OPAC is a reflection of the health of the system. The OPAC should be spilling out onto our website and beyond, mashing it up with other sites. The only way that can happen is with a rich API, which we don’t have.

The title of systems librarian is becoming redundant because we all have a responsibility and role in maintaining the health of library systems. In today’s information ecology, there is no destination — we’re online experiencing information everywhere.

There is no way to predict how the information ecology will change, so we need systems that will be flexible and can grow and change over time. (Sopac 2.0 will be released later this year for libraries who want to do something different with their OPACs.) Containers will fail. Containers are temporary. We cannot hang our hat on one specific format — we need systems that permit portability of data.

Nobody in libraries talks about “the enterprise” like they do in the corporate world. Design and development of the enterprise cannot be done by a committee, unless they are simply advisors.

The 21st century library remains un-designed – so let’s get going on it.

CiL 2008: Libraries A-Twitter and Using del.icio.us

Speakers: Aysegul Kapucu, Athena Hoeppner, and Doug Dunlop (University of Central Florida)

del.icio.us is a free social bookmarking tool that can be organized with tags and bundles. UCF wanted to see if they could increase access points for library resources with on-the-fly lists for classes and individuals.

They loaded all of their databases with EZProxy string pre-pended to the URL. Then they tagged them.

The del.icio.us browser buttons were installed on the reference desk. During the reference interview, they tagged resources, and at the end, they would give the user a link to all the resources that were tagged for them. For classes, they tag the bookmarks with the course short code and include the resources listed by the professor in their syllabus. Two topical accounts are being developed through a collaboration with faculty and graduate students in Art and Engineering.

They surveyed 300+ faculty and students and received 50 responses, most of which came from seniors and reflected the courses that were included in the tagging project. 70% of the respondents had not used del.icio.us prior to the library’s involvement, which is probably due to the relatively small number of users as compared to other social networking tools like Facebook.

I could see del.icio.us being used as a replacement for hand coded subject guides or commercial products that do the same. Since it’s easy to add or edit on the fly, the guides could be more relevant than static lists.


Speakers: Michael Sauers and Christa Burns

Twitter is microblogging, like status updates on MySpace and Facebook. It’s like instant messaging, but it is asynchronous. Twitter is experiential — you have to do it with people you know to get it.

All of the twitterers in the room were wetting themselves with the excitement of getting to twitter about a Twitter presentation.

Libraries can use Twitter to broadcast information about what is going on at the library. At the Nebraska Library Center, the reference librarians send out tweets of the questions they get (not the answers). A few cities have traffic and weather reports sent out via Twitter. “We can’t get enough information about weather. Especially catalogers who don’t have windows.”

Twitter is ephemeral.

7 Tips To a Good Twitter Experience from Meryl is a good resource for new twitterers.

They must put the “Twitter is like…” slide presentation somewhere everyone can see it.

CiL 2008: What Do Users Really Do in Their Native Habitat?

Speakers: Pascal Lupien and Randy Oldham

Unsubstantiated assumptions about Millennials cause libraries to make poor choices in providing services and resources. Lupien and Oldham spent some time studying how students actually use the tools we think they use. They used typical survey and focus group methodologies, which make for rather boring presentation recaps, so I won’t mention them.

Study found that only 9% of students used PDAs, and tended to be among older students. 69% of students had cell phones, but only 17% of them have ever used them to browse the Internet. 93% of student have used a chat client, and most have used them for academic purposes several times per week. 50% of users had never used online social network applications for academic group work.

The focus groups found that students preferred email over online social networks for group work. Students are more willing to share the results of their work with friends than with other classmates.

42% of students has never played online games, and men were three times more likely to do so than women. Only 4.1% were involved with online virtual worlds like World of Warcraft and Second Life.

The survey respondents indicated they were more likely to go to the library’s website first rather than Google. The focus groups also confirmed this, in addition to indicating that the library had the best sources of information despite being the most difficult to manage.

Students are reluctant to mix personal and academic computing. The uptake on online social networks for academic use has been slow, but will likely increase, and we have to ask, “is this the best use of our resources and time?” Our priorities need to be more on improving the services we already offer, such as our websites and search tools. “Rather than looking at technologies & trying to find a use for them in our environment, we should determine what our students need & seek solutions to meet those needs.”


Speaker: John Law

Proquest conducted a survey of seven universities across North America and the United Kingdom, involving 60 students. As with Lupien and Oldham’s study, they conducted it anonymously. Observations were conducted in a variety of locations, from the library to dorm rooms. They used a program like web conferencing software to capture the remote sessions.

Law gave an anecdote of a fourth year student who did all the things librarians want students to do when doing research, and when he was asked why, the student gave all the right answers. Then, when he was asked how long he had been doing his research that way, he indicated something like six weeks, after a librarian had come to his class to teach them about using the library’s resources. Library instruction works.

Course instructors are also influential. “My English instructor told me to use JSTOR.”

Brand recognition is fine, but it doesn’t necessarily effect the likelihood that resources will be used more or less.

Students use abstracts to identify relevant articles, even when the full text is available. They’re comfortable navigating in several different search engines, but not as well with library websites in locating relevant resources. Users don’t always understand what the search box is searching (books, articles, etc.), and can find it to be discouraging. A-Z databases page is too unmanageable for most users, particularly when starting their research.

Students are using Google for their research, but mainly for handy look-ups and not as a primary research tool. Those who use Google as a primary research tool do so because they aren’t as concerned with quality or are insufficiently aware of library eresources or have had bad experiences with library eresources.

Librarians, students use Google and Wikipedia the same way you do. (We know you all use those tools, so don’t even try to deny it.)

Students laughed at surveyors when asked how they use online social networks for academic purposes.

CiL 2008: Widgets, Tools, & Doodads for Library Webmasters

Speakers: Darlene Fitcher & Frank Cervone

Firefox:

  • SafeCache – prevents applications from accessing your browser cache
  • SafeHistory -prevents applications from viewing browser history
  • FoxMarks – automatically synchronizes bookmarks
  • Firefox Environment Backup Extension – synchronize between computers, including other preference settings

Other web tools for collaboration:

  • Meebo Chat Widget – IM chat for reference; good on the null search page of a OPAC and everywhere else
  • LinkBunch –  put multiple links into one small link
  • DocSyncer – sync docs on local machine to Google Docs (this means everything)
  • twhirl – microblogging or keeping tabs on what colleagues are doing
  • polldaddy – put a poll on your website
  • VisCheck – check the colors on your website for various color blindness compatibilities
  • Feng GUI – alternate to eye-tracking
  • Browsershots – creates screenshots in different browsers
  • Flickr Photo badge – photos from your library’s photo stream
  • Photoshop Express – free online photo editor
  • AddThis – button that allows users to bookmark or share pages; also provides stats on use
  • Google Countdown – widget for announcing events
  • Altavista Bablefish site translator – translate your pages on the fly
  • ProcessTamer – keep tabs on processes that bog down your computer
  • File Hamster – keep track of versions of files w/out having to use MS
  • Syncback Freeware – backup files, scheduler for automating, synchronization, machine to machine via ftp
  • LinkExtractor
  • reCaptcha – “stop spam. read books.” using captcha to clean up OCR scanning
  • Anonymouse – check if resources are secure outside of IP range or for privacy
  • Prism – use web applications as a local client

More!

Fun:

  • LOLinator

[will add missing URLs later when I have more reliable connection]

CiL 2008: What’s New With Federated Search

Speakers: Frank Cervone & Jeff Wisniewski

Cervone gave a brief over-view of federated searching, with Wisniewski giving a demonstration of how it works in the real world (aka University of Pittsburgh library) using WebFeat. UofP library has a basic search front and center on their home page, and then a more advanced searching option under Find Articles. They don’t have a Database A-Z list because users either don’t know what database means in this context or can’t pick from the hundreds available.

Cervone demonstrated the trends in using meta search, which seems to go up and down, but over-all is going up. The cyclical aspect due to quarter terms was fascinating to see — more dramatic than what one might find with semester terms. Searches go up towards mid-terms and finals, then drop back down afterwards.

According to a College & Research Libraries article from November 2007, federated search results were not much different from native database searches. It also found that faculty rated results of federated searching much higher than librarians, which begs the question, “Who are we trying to satisfy — faculty/students or librarians.”

Part of why librarians are still unconvinced is because vendors are shooting themselves in the foot in the way they try to sell their products. Yes, federated search tools cannot search all possible databases, but our users are only concerned that they search the relevant databases that they need. De-duplication is virtually impossible and depends on the quality of the source data. There are other ways that vendors promote their products in ways that can be refuted, but the presenters didn’t spend much time on them.

The relationships between products and vendors is incestuous, and the options for federated searching are decreasing. There are a few open source options, though: LibraryFind, dbWiz, Masterkey, and Open Translators (provides connectors to databases, but you have to create the interface). Part of why open source options are being developed is because commercial vendors aren’t responding quickly to library needs.

LibraryFind has a two-click find workflow, making it quicker to get to the full-text. It also can index local collections, which would be handy for libraries who are going local.

dbWiz is a part of a larger ERM tool. It has an older, clunkier interface than LibraryFind. It doesn’t merge the results.

Masterkey can search 100 databases at a time, processing and returning hits at the rate of 2000 records per second, de-duped (as much as it can) and ranked by relevance. It can also do faceted browsing by library-defined elements. The interface can be as simple or complicated as you want it to be.

Federated searching as a stand-alone product is becoming passe as new products for interfacing with the OPAC are being developed, which can incorporate other library databases. vufind, WorldCat local, Encore, Primo, and Aquabrowser are just a few of the tools available. NextGen library interfaces aim to bring all library content together. However, they don’t integrate article-level information with the items in your catalog and local collections very well.

Side note: Microsoft Enterprise Search is doing a bit more than Google in integrating a wide range of information sources.

Trends: Choices from vendors is rapidly shrinking. Some progress in standards implementation. Visual search (like Grokker) is increasingly being used. Some movement to more holistic content discovery. Commercial products are becoming more affordable, making them available to institutions of all sizes of budgets.

Federated Search Blog for vendor-neutral info, if you’re interested.

CiL 2008: Going Local in the Library

Speaker: Charles Lyons

[Speakers for track C will all be introduced in haiku form.]

Local community information has been slower to move online than more global information provided by sources such as search engines and directories, but that is changing. Google can provide directory information, but they can’t tell you which of the barbers listed near you, for example, are good ones. They’re trying, but it’s much harder to gather and index that information. “In ur community, inforimin ur localz.” The local web can be something deeper, hyper, semantic.

Local information can sound boring if it doesn’t effect you directly. For example, information about street repairs can be important if it is happening along the routes you drive regularly. The local web connects the real world and the virtual world. The local web is bringing a sense of place to the Internet.

Libraries provide access to local information such as genealogy, local history, local government info, etc.

Local search engines started off as digital phone books, but now they are becoming more integrated with additional information such as maps and users reviews. Ask.com provides walking directions as well as driving directions, which I did not know but plan to make use of in the future. By using tools like Google Custom Search, libraries are creating local search engines, rather than just having a page of local links. MyCommunityInfo.ca is a popular search engine for residents of London, Ontario.

Local blogs also provide information about communities, so creating a local blog directory might be useful. Be sure to add them to your local search engine. Local news sites blend user-generated information with traditionally published sources. Useful news sites will allow users to personalize them and add content. Libraries are involved in creating local online communities (see Hamilton Public Library in Ontario).

Local data is being aggregated by sites like EveryBlock, which pulls information from the deep web. It’s currently available in three cities (Chicago, New York, & San Francisco) with plans for expansion, and once the grant ends, the code will be opened to anyone.

Wikipedia is a start for providing local information, and a few libraries are using wiki technology to gather more detailed local information.

Metadata such as geotagging allows more automation for gathering locally relevant information. Flickr provides geographic feeds, which can be streamed on local information sites.

Libraries are using Google Maps mashups to indicate their locations, but could be doing more with it. Libraries could create maps of historical locations and link them to relevant information sites.

No successful revenue generation has been formulated for local information sites. Most local sites are generated by passionate individuals. Libraries, which are not revenue generating sources anyway, are better poised to take on the responsibility of aggregating and generating local information. Particularly since we’re already in the role of information provision.

Libraries can be the lense into local information.

css.php