library use of web-based research guides jimmy ghaphery and erin white information technology and libraries | march abstract this paper describes the ways in which libraries are currently implementing and managing web- based research guides (a.k.a. pathfinders, libguides, subject guides, etc.) by examining two sets of data from the spring of . one set of data was compiled by visiting the websites of ninety-nine american university arl libraries and recording the characteristics of each site’s research guides. the other set of data is based on an online survey of librarians about the ways in which their libraries implement and maintain research guides. in conclusion, a discussion follows that includes implications for the library technology community. selected literature review while there has been significant research on library research guides, there has not been a recent survey either of the overall landscape or of librarian attitudes and practices. there has been recent work on the efficacy of research guides as well as strategies for their promotion. there is still work to be done on developing a strong return on investment metric for research guides, although the same could probably be said for other library technologies including websites, digital collections, and institutional repositories. subject-based research guides have a long history in libraries that predates the web as a service- delivery mechanism. a literature-review article from found that research on the subject gained momentum around with the advent of electronic research guides, and that there was a need for more user-centric testing. by the mid- s, it was rare to find a library that did not offer research guides through its website. the format of guides has certainly shifted over time to database-driven efforts through local library programming and commercial offerings. a number of other articles start to answer some of the questions about usability posed in the literature review by vileno. in , grays, del bosque, and costello used virtual focus groups as a test bed for guide evaluation. two articles from the august issue of the journal of library administration contain excellent literature reviews and look toward marketing, assessment, and best practices. also in , vileno followed up on the literature review with usability testing that pointed toward a number of areas in which users experienced difficulties with research guides. jimmy ghaphery (jghapher@vcu.edu) is head, library information systems and erin white (erwhite@vcu.edu) is web systems librarian, virginia commonwealth university libraries, richmond, va. mailto:jghapher@vcu.edu library use of web-based research guides | ghaphery and white in terms of cross-library studies, an interesting collaboration in between cornell and princeton universities found that students, faculty, and librarians perceived value in research guides, but that their qualitative comments and content analysis of the guides themselves indicated a need for more compelling and effective features. the work of morris and grimes from should also be mentioned; the authors surveyed university libraries, finding that it was rare to find a library with formal management policies for their research guides. most recently, libguides has emerged as a leader in this arena, offering a popular software-as-a- service (saas) model and as such is not yet heavily represented in the literature. a multichapter libguides lita guide is pending publication and will cover such topics as implementing and managing libguides, setting standards for training and design, and creating and managing guides. arl guides landscape during the week of march rd, , the authors visited the websites of american university arl libraries to determine the prevalence and general characteristics of their subject-based research guides. in general, the visits reinforced the overarching theme within the literature that subject-based research guides are a core component of academic library web services. all libraries offered research guides that were easy to find from the library home page. libguides was very prominent as a platform, in production at of the libraries. among these, it appeared that at least libraries were in the process of migrating from a previous system (either a homegrown, database-driven site or static html pages) to libguides. in addition to the presence and platform, the authors recorded additional information about the scope and breadth of each site’s research guides. for each site, the presence of course-based research guides was recorded. in some cases the course guides had a separate listing, whereas in others they were intermingled with the subject-based research guides. course guides were found on of the libraries visited. of these, were also libguides sites. it is certainly possible that course guides are being deployed at some of the other libraries but were not immediately visible in visiting the websites, or that course guides may be deployed through a course management system. nonetheless, it appears that the use of libguides encourages the presence of public-facing course guides. qualitatively, there was wide diversity of how course guides were organized and presented, varying from a simple a-to-z listing of all guides to separately curated landing pages specifically organized by discipline. the number of guides was recorded for each libguides site. it was possible to append “/browse.php?o=a” to the base url to determine how many guides and authors were published at each site. this php extension was the publicly available listing of all guides on each libguides platform. the “/browse.php?o=a” extension no longer publicly reports these statistics; however, findings could be reproduced by manually counting the number of guides and authors on each site. the authors confirmed the validity of this method in the fall of by revisiting four sites and finding that the numbers derived from manual counting were in line with the previous findings. of information technology and libraries | march the libguides sites we observed, a total of , guides were counted from , authors for an average of guides per author. on average, each site had guides from authors (median of guides; authors). at the high end of the scale, one site had guides from authors. based on the volume observed, libraries appear to be investing significant time toward the creation, and presumably the maintenance, of this content. in addition to creation and ongoing maintenance, such long lists of topics raise a number of usability issues that libraries will also be wise to keep in mind. survey the literature review and website visits call out two strong trends: . research guides are as commonplace as books in libraries, . libguides is the elephant in the room, so much so that it is hard to discuss research guides without discussing libguides. based on preliminary findings from the literature review and survey, we looked to further describe how libraries are supporting, innovating, implementing, and evaluating their research guides. a ten-question survey was designed to better understand how research guides sit within the cultural environment of libraries. it was distributed to a number of professional discussion lists the week of april , (see appendix). the following lists were used in an attempt to get a balance of opinion from populations of both technical and public services librarians: code lib, web lib, lita-l, lib-ref-l, and ili-l. the survey was made available for two weeks following the list announcements. survey response was very strong, with responses ( libraries) received without the benefit of any follow-up recruitment. ten institutions submitted more than one response. in these cases only the first response was included for analysis. we did not complete a response for our own institution. the vast majority ( , %) of respondents were from college or university libraries. of the remaining , ( %) were from community college libraries, with only ( %) identifying themselves as public, school, private, or governmental. among the college and university libraries, ( %) identified themselves as members of the arl, which comprises members. in terms of “what system best describes your research guides by subject?” the results were similar to the survey of arl websites. most libraries ( , %) reported libguides as their system, followed by “customized open source system” and “static html pages,” both at responses ( % each). sixteen libraries ( %) reported using a homegrown system, with three libraries ( %) reporting “other commercial system.” in terms of initiating and maintaining a guides system, much of the work within libraries seems to be happening outside of library systems departments. when asked which statement best described who selected the guides system, respondents ( %) indicated their library research library use of web-based research guides | ghaphery and white guides were “initiated by public services,” followed closely by “more of a library-wide initiative” at responses ( %). in the middle at responses ( %) was “initiated by an informal cross- departmental group.” only respondents ( %) selected “initiated by systems,” with the top down approach of “initiated by administration” gathering responses ( %). when narrowing the responses to those sites that are using libguides or campus guides, the portrait is not terribly different, with % library-wide, % public services, % informal cross-departmental, % administration, and systems trailing at %. likewise there was not a strong indication of library systems involvement in maintaining or supporting research guides. sixty-nine responses ( %) indicated “no ongoing involvement” and an additional ( %) indicated “n/a we do not have a systems department.” there were only responses ( %) stating “considerable ongoing involvement,” with the balance of responses ( %) for “some ongoing involvement.” not surprisingly, there was a correlation between the type of research guide and the amount of systems involvement. for sites running a “customized open source system,” “other commercial system,” or “homegrown system,” at least % of responses indicated either “considerable” or “some” ongoing systems involvement. in contrast, % of sites running libguides or campusguides indicated “considerable” or “some” technical involvement. further, the libguides and campusguides users recorded the highest percentage ( %) of “no ongoing involvement” compared to % of all respondents. interestingly, % of libguides and campus guides users answered “n/a we do not have a systems department,” which is not significantly higher than all respondents for this question at %. the level of interaction between research guides and enterprise library systems was not reported as strong. when asked “which statement best describes the relationship between your web content management system and your research guides?” responses ( %) indicated that “our content management system is independent of our research guides” with an additional responses ( %) indicating that they did not have a content management system (cms). only respondents ( %) said that their cms was integrated with their research guides with a remaining ( %) saying that their cms was used for “both our website and our research guides.” a similar portrait was found in seeking out the relationship between research guides and discovery/federated search tools. when asked “which statement best describes the relationship between your discovery/federated search tool and your research guides?” roughly half of the respondents ( , %) did not have a discovery system (“n/a we do not have a discovery tool”). only respondents ( %) selected “we prominently feature our discovery tool on our guides,” whereas more than double that number, ( %), said “we typically do not include our discovery tool on our guides.” fifty four respondents ( %) took the middle path of “our discovery tool is one of many search options we feature on our guides.” in the case of both discovery systems and content management systems, it seems that research guides are typically not deeply integrated. when asked “what other type of content do you host on your research guides system?” respondents selected from a list of choices as reflected in table . information technology and libraries | march answer total percent libguides/campusguides course pages % % “how to” instruction % % alphabetical list of all databases % % “about the library” information (for example hours, directions, staff directory, event) % % digital collections % % everything—we use the research guide platform as our website % % none of the above % % table . other types of content hosted on research guides system these answers reinforce the portrait of integration within the larger library web presence. while the research guides platform is an important part of that presence, significant content is also being managed by libraries through other systems. it is also consistent with the findings from the arl website visits, where course pages were consistently found within the research guides platform. for sites reporting libguides or campusguides as their platform, inclusion of course pages and how-to instruction was even higher, at % and %, respectively. another multi-answer question sought to determine what types of policies are being used by libraries for the management of research guides: “which of the following procedures or policies do you have in place for your research guides?” responses are summarized in table . library use of web-based research guides | ghaphery and white answer total percent percent using libguides/campusguides style guides for consistent presentation maintenance and upkeep of guides link checking required elements such as contact information, chat, pictures, etc. training for guide creators transfer of guides to another author due to separation or change in duties defined scope of appropriate content allowing and/or moderating user tags, comments, ratings none of the above controlled vocabulary/tagging system for managing guides table . management policies/procedures for research guides while nearly one in five libraries reported none of the policies in place at all, the responses indicate that there is effort being applied toward the management of these systems. the highest percentage for any given policy was % for “style guides for consistent presentation.” best practices in these areas could be emerging or many of these policies could be specific to individual library needs. as with the survey question on content, the research-guides platform also has a role with the libguides and campusguides users reporting much higher rates of policies for “controlled vocabulary/tagging” ( % vs. %) and “required elements” ( % vs. %). in both information technology and libraries | march of these cases, it is likely that the need for policies arise from the availability of these features and options that may not be present in other systems. based on this supposition, it is somewhat surprising that the libguides and campusguides sites reported the same lack of policy adoption (none of the above; %). the final question in the survey further explored the management posture for research guides by asking a free-text question: “how do you evaluate the success or failure of your research guides?” results were compiled into a spreadsheet. the authors used inductive coding to find themes and perform a basic data analysis on the responses, including a tally of which evaluation methods were used and how often. one in five institutions ( respondents, . %) looked only to usage stats, while seven respondents ( %) indicated that their library had performed usability testing as part of the evaluation. forty-our respondents ( . %) said they had no evaluation method in place (“ouch! it hurts to write that.”), though many expressed an interest or plans to begin evaluation. another emerging theme included ten respondents who quantified success in terms of library adoption and ease of use. this included one respondent who had adopted libguides in light of prohibitive it regulations (“we choose libguides because it would not allow us to create class specific research webpages”). several institutions also expressed frustration with the survey instrument because they were in the process of moving from one guides system to another and were not sure how to address many questions. most responses indicated that there are more questions than answers regarding the efficacy of their research guides, though the general sentiment toward the idea of guides was positive, with words such as “positive,” “easy,” “like,” and “love” appearing in responses. countering that, respondents indicated that their libraries’ research-guides projects had fallen through. conclusion this study confirms previous research that web-based research guides are a common offering, especially in academic libraries. adding to this, we have recorded a quantitative adoption of libguides both through visiting arl websites and through a survey distributed to library listservs. further, this study did not find a consistent management or assessment practice for library research guides. perhaps the most interesting finding from this study is the role of library systems departments with regard to research guides. it appears that many library systems departments are not actively involved in either the initiation or ongoing support of web-based research guides. what are the implications for the library technology community and what questions arise for future research? the apparent ascendancy of libguides over local solutions is certainly worth considering and in part demonstrates some comfort within libraries for cloud computing and saas. time will tell how this might spread to other library systems. the popularity of libguides, at its heart a specialized content management system, also calls into question the vitality and adaptability of local content management system implementations in libraries. more generally, does the desire to professionally select and steward information for users on research guides indicate librarian misgivings about the usability of enterprise library systems? how do attitudes library use of web-based research guides | ghaphery and white toward research guides differ between public services and technical services? hopefully these questions serve as a call for continued technical engagement with library research guides. what shape that engagement may have in the future is an open question, but based on the prevalence and descriptions of current implementations, such consideration by the library technology community is worthwhile. references . luigina vileno, “from paper to electronic, the evolution of pathfinders: a review of the literature,” reference services review , no. ( ): – . . martin courtois, martha higgins, aditya kapur, “was this guide helpful? users’ perceptions of subject guides,” reference services review , no. ( ): – . . lateka j. grays, darcy del bosque, and kristen costello, “building a better m.i.c.e. trap: using virtual focus groups to assess subject guides for distance education students,” journal of library administration , no. / ( ): – . . mira foster et al., “marketing research guides: an online experiment with libguides,” journal of library administration , no. / (july/september, ): – ; alisa c. gonzalez and theresa westbrock, “reaching out with libguides: establishing a working set of best practices,” journal of library administration , no. / (july/september, ): – . . luigina vileno, “testing the usability of two online research guides,” partnership: the canadian journal of library and information practice and research , no. ( ), http://journal.lib.uoguelph.ca/index.php/perj/article/view/ (accessed august , ). . angela horne and steve adams, “do the outcomes justify the buzz? an assessment of libguides at cornell university and princeton university—presentation transcript,” presented at the association of academic and research libraries, seattle, wa, , http://www.slideshare.net/smadams/do-the-outcomes-justify-the-buzz-an-assessment-of- libguides-at-cornell-university-and-princeton-university (accessed august , ). . sarah morris and marybeth grimes, “a great deal of time and effort: an overview of creating and maintaining internet-based subject guides,” library computing , no. ( ): – . . mathew miles and scott bergstrom, “classification of library resources by subject on the library website: is there an optimal number of subject labels?” information technology & libraries , no. (march ): – , http://www.ala.org/lita/ital/files/ / /miles.pdf (accessed august , ). . association of research libraries, “association of research libraries: member libraries,” http://www.arl.org/arl/membership/members.shtml (accessed october , ). http://journal.lib.uoguelph.ca/index.php/perj/article/view/ http://www.slideshare.net/smadams/do-the-outcomes-justify-the-buzz-an-assessment-of-libguides-at-cornell-university-and-princeton-university http://www.slideshare.net/smadams/do-the-outcomes-justify-the-buzz-an-assessment-of-libguides-at-cornell-university-and-princeton-university http://www.ala.org/lita/ital/files/ / /miles.pdf http://www.arl.org/arl/membership/members.shtml information technology and libraries | march appendix. survey library use of web-based research guides please complete the survey below. we are researching libraries’ use of web-based research guides. please consider filling out the following survey, or forwarding this survey to the person in your library who would be in the best position to describe your library’s research guides. responses are anonymous. thank you for your help! jimmy ghaphery, vcu libraries erin white, vcu libraries ) what is the name of your organization? __________________________________ note that the name of your organization will only be used to make sure multiple responses from the same organization are not received. any publication of results will not include specific names of organizations. ) which choice best describes your library? o arl o university library o college library o community college library o public library o school library o private library o governmental library o nonprofit library ) what type of system best describes your research guides by subject? o libguides or campusguides o customized open source system o other commercial system o homegrown system o static html pages ) which statement best describes the selection of your current research guides system? o initiated by administration o initiated by systems o initiated by public services o initiated by an informal cross-departmental group o more of a library-wide initiative library use of web-based research guides | ghaphery and white ) how much ongoing involvement does your systems department have with the management of your research guides? o no ongoing involvement o some ongoing involvement o considerable ongoing involvement o n/a we do not have a systems department ) what other type of content do you host on your research guides system? o course pages o “how to” instruction o alphabetical list of all databases o “about the library” information (for example: hours, directions, staff directory, events) o digital collections o everything—we use the research guide platform as our website o none of the above ) which statement best describes the relationship between your discovery/federated search tool and your research guides? o we typically do not include our discovery tool on our guides o our discovery tool is one of many search options we promote on our guides o we prominently feature our discovery tool on our guides o n/a we do not have a discovery tool ) which statement best describes the relationship between your web content management system and your research guides? o our content management system is independent of our research guides o our content management system is integrated with our research guides o our content management system is used for both our website and our research guides o n/a we do not have a content management system ) which of the following procedures or policies do you have in place for your research guides? o defined scope of appropriate content o required elements such as contact information, chat, pictures, etc. o style guides for consistent presentation o allowing and/or moderating user tags, comments, ratings o training for guide creators o controlled vocabulary/tagging system for managing guides o maintenance and upkeep of guides o link checking information technology and libraries | march o transfer of guides to another author due to separation or change in duties o none of the above ) how do you evaluate the success or failure of your research guides? [free text] usability test results for a discovery tool in an academic library jody condit fagan meris mandernach carl s. nelson jonathan r. paulo grover saunders information technology and libraries | march abstract discovery tools are emerging in libraries. these tools offer library patrons the ability to concurrently search the library catalog and journal articles. while vendors rush to provide feature-rich interfaces and access to as much content as possible, librarians wonder about the usefulness of these tools to library patrons. to learn about both the utility and usability of ebsco discovery service, james madison university (jmu) conducted a usability test with eight students and two faculty members. the test consisted of nine tasks focused on common patron requests or related to the utility of specific discovery tool features. software recorded participants’ actions and time on task, human observers judged the success of each task, and a post–survey questionnaire gathered qualitative feedback and comments from the participants. participants were successful at most tasks, but specific usability problems suggested some interface changes for both ebsco discovery service and jmu’s customizations of the tool. the study also raised several questions for libraries above and beyond any specific discovery-tool interface, including the scope and purpose of a discovery tool versus other library systems, working with the large result sets made possible by discovery tools, and navigation between the tool and other library services and resources. this article will be of interest to those who are investigating discovery tools, selecting products, integrating discovery tools into a library web presence, or performing evaluations of similar systems. introduction discovery tools appeared on the library scene shortly after the arrival of next-generation catalogs. the authors of this paper define discovery tools as web software that searches journal-article and library-catalog metadata in a unified index and presents search results in a single interface. this differs from federated search software, which searches multiple databases and aggregates the results. examples of discovery tools include serials solutions summon, ebsco discovery service, jody condit fagan (faganjc@jmu.edu) is director, scholarly content systems, meris mandernach (manderma@jmu.edu) is collection management librarian, carl s. nelson (nelsoncs@jmu.edu) is digital user experience specialist, jonathan r. paulo (paulojr@jmu.edu) is education librarian, and grover saunders (saundebn@jmu.edu) is web media developer, carrier library, james madison university, harrisonburg, va. mailto:faganjc@jmu.edu mailto:manderma@jmu.edu mailto:nelsoncs@jmu.edu mailto:paulojr@jmu.edu mailto:saundebn@jmu.edu usability test results for a discovery tool in an academic library | fagan et al ex libris primo, and oclc worldcat local; examples of federated search software include serials solutions webfeat and ebsco integrated search. with federated search software, results rely on the search algorithm and relevance ranking as well as each tool’s algorithms and relevance rankings. discovery tools, which import metadata into one index, apply one set of search algorithms to retrieve and rank results. this difference is important because it contributes to a fundamentally different user experience in terms of speed, relevance, and ability to interact consistently with results. combining the library catalog, article indexes, and other source types in a unified interface is a big change for users because they no longer need to choose a specific search tool to begin their search. research has shown that such a choice has long been in conflict with users’ expectati ons. federated search software was unable to completely fulfill users’ expectations because of its limited technology. now that discovery tools provide a truly integrated search experience, with greatly improved relevance rankings, response times, and increased consistency, libraries can finally begin to meet this area of user expectation. however, discovery tools present new challenges for users: will they be able to differentiate between source types in the integrated results sets? will they be able to limit large results sets effectively? do they understand the scope of the tool and that other online resources exist outside the tool’s boundaries? the sea change brought by discovery tools also raises challenges for librarians, who have grown comfortable with the separation between the library catalog and other online databases. discovery tools may mask important differences between disciplinary searching, and they do not currently offer discipline-specific strategies or limits. they also lack authority control, which makes topical precision a challenge. their usual prominence on library websites may direct traffic away from carefully cultivated and organized collections of online resources. discovery tools offer both opportunities and challenges for library instruction, depending on the academic discipline, users’ knowledge, and information-seeking need. james madison university (jmu) is a predominantly undergraduate institution of approximately , students in virginia. jmu has a strong information literacy program integrated into the curriculum through the university’s information seeking skills test (isst). the isst is completed before students are able to register for third-semester courses. additionally, the library provides an information literacy tutorial, “go for the gold,” that supports the skills needed for the isst. jmu launched ebsco discovery service (eds) in august after participating as a beta development partner in spring and summer . as with other discovery tools, the predominant feature of eds is integration of the library catalog with article databases and other types of sources. at the time of this study, eds had a few differentiating features. first, because of ebsco’s business as a database and journal provider, article metadata was drawn from a combination of journal-publisher information and abstracts and index records. the latter included robust subject indexing (e.g., the medical subject headings in cinahl). the content searched by eds varies by information technology and libraries | march institution according to the institution’s subscription. jmu had a large number of ebsco databases and third-party database subscriptions through ebsco, so the quantity of information searched by eds at jmu is quite large. eds also allowed for extensive customization of the tool, including header navigation links, results-screen layout, and the inclusion of widgets in the right-hand column of the results screen. jmu libraries developed a custom “quick search” widget based on eds for the library home page (see figure ), which allows users to add limits to the discovery-tool search and assists with local authentication requirements. based on experience with a pilot test of the open-source vufind next-generation catalog, jmu libraries believed users would find the ability to limit up-front useful, so quick search’s first drop-down menu contained keyword, title, and author field limits; the second drop-down contained limits for books, articles, scholarly articles, “just leo library catalog,” and the library website (which did not use eds). the “just leo library catalog” option limited the user’s search to the library catalog database records but used the eds interface to perform the search. to access the native catalog interface, a link to leo library catalog was included immediately above the search box as well as in the library website header. figure . quick search widget on jmu library homepage usability test results for a discovery tool in an academic library | fagan et al evaluation was included as part of the implementation process for the discovery t ool, and therefore a usability test was conducted in october . the purpose of the study was to explore how patrons used the discovery tool, to uncover any usability issues with the chosen system and to investigate user satisfaction. specific tasks addressed the use of facets within the discovery tool, patrons’ use of date limiters, and the usability of the quick search widget. the usability test also had tasks in which users were asked to locate books and articles using only the discovery tool, then repeat the task using anything but the discovery tool. this article interprets the usability study’s results in the context of other local usability tests and web-usage data from the first semester of use. some findings were used to implement changes to quick search and the library website, and to recommend changes to ebsco; however, other findings suggested general questions related to discovery tool software that libraries will need to investigate further. literature review literature reviewed for this article included some background reading on users and library catalogs, library responses to users’ expectations, usability studies in libraries, and usability studies of discovery tools specifically. the first group of articles comprised a discussion about the limitations of traditional library catalogs. the strengths and weaknesses of library catalogs were reported in several academic libraries’ usability studies. calhoun recognized that library users’ preference for google caused a decline in the use and value of library catalogs, and encouraged library leaders to “establish the catalog within the framework of online information discovery systems.” this awareness of changes in user expectations during a time when google set the benchmark for search simplicity was echoed by numerous authors who recognized the limits of library catalogs and expressed a need for the catalog to be greatly modernized to keep pace with the evolution of the web. libraries have responded in several ways to the call for modernization, most notably through investigations related to federated searching and next-generation catalogs. several articles have presented usability studies results for various federated searching products. fagan provided a thorough literature review of faceted browsing and next-generation catalogs. western michigan university presented usability study results for the next-generation catalog vufind, revealing that participants took advantage of the simple search box but did not use the next-generation catalog features of tagging, comments, favorites, and sms texting. the university of minnesota conducted two usability studies of primo and reported that participants were satisfied with using primo to find known print items, limit by author and date, and find a journal title. tod olson conducted a study with graduate students and faculty using the aquabrowser interface, and his participants located sources for their research they had not previously been able to find. information technology and libraries | march the literature also revealed both opportunities and limitations of federated searching and next- generation catalogs. allison presented statistics from google analytics for an implementation of encore at the university of nebraska-lincoln. the usage statistics revealed an increased use of article databases as well as an increased use of narrowing facets such as format and media type, and library location. allison concluded that encore increased users’ exposure to the entire collection. breeding concluded that federated searching had various limitations, especially search speed and interface design, and was thus unable to compete with google scholar. usability studies of next-generation catalogs revealed a lack of features necessary to fully incorporate an entire library’s collection. breeding also recognized the limitations of next-generation library catalogs and saw discovery tools as their next step in evolution: “it’s all about helping users discover library content in all formats, regardless of whether it resides within the physical library or among its collections of electronic content, spanning both locally owned materials and those accessed remotely through subscriptions.” the dominant literature related to discovery tools discussed features, reviewed them from a library selector perspective, summarized academic libraries’ decisions following selection, presented questions related to evaluation after selection, and offered a thorough evaluation of common features. allison concluded that “usability testing will help clarify what aspects need improvement, what additions will make [the interface] more useful, and how the interface can be made so intuitive that user training is not needed.” breeding noted “it will only be through the experience of library users that these products will either prove themselves or not.” libraries have been adapting techniques from the field of usability testing for over a decade to learn more about user behavior, usability, and user satisfaction, with library web sites and systems. rubin and chisnell and dumas and redish provided an authoritative overview of the benefits and best practices of usability testing. in addition, campbell and norlin and winters offered specific usability methodologies for libraries. worldcat local has dominated usability studies of discovery tools published to date. ward, shadle, and mofield conducted a usability study at the university of washington. although the second round of testing was not published, the first round involved seven undergraduate and three graduate students; its purpose “was to determine how successful uw students would be in using worldcat local to discover and obtain books and journal articles (in both print and electronic form) from the uw collection, from the summit consortium, and from other worldcat libraries.” although participants were successful at completing these tasks, a few issues arose out of the usability study. users had difficulty with the brief item display because reviews were listed higher than the actual items. the detailed item display also hindered users’ ability to decipher between various editions and formats. the second round of usability testing, not yet published, included tasks related to finding materials on specific subject areas. usability test results for a discovery tool in an academic library | fagan et al boock, chadwell, and reese conducted a usability study of worldcat local at oregon state university. the study included four tasks and five evaluative questions. forty undergraduate students, sixteen graduate students, twenty-four library employees, four instructors, and eighteen faculty members took part in the study. they summarized that users found known-title searching to be easier in the library catalog but found topical searches to be more effective in worldcat local.the participants preferred worldcat local for the ability to find articles and search for materials in other institutions. western washington university also conducted a usability study of worldcat local. they selected twenty-four participants with a wide range of academic experience to conduct twenty tasks in both worldcat local and the traditional library catalog. the comparison revealed several problems in using worldcat local, including users’ inability to determine the scope of the content, confusion over the intermixing of formats, problems with the display of facet option, and difficulty with known-item searches. western washington university decided not to implement worldcat local. oclc published a thorough summary of several usability studies conducted mostly with academic libraries piloting the tool, including the university of washington; the university of california (berkeley, davis, and irvine campuses); ohio state university; the peninsula library system in san mateo, california; and the free library of urbana and the des plaines public library, both in illinois. the report conveyed favorable user interest in searching local, group, and global collections together. users also appreciated the ability to search articles and books together. the authors commented, “however, most academic participants in one test (nine of fourteen) wrongly assumed that journal article coverage includes all the licensed content available at their campuses.” oclc used the testing results to improve the order of search results, provide clarity about various editions, improve facets for narrowing a search, provide links to electronic resources, and increase visibility of search terms. at grand valley state university, doug way conducted an analysis of usage statistics after implementing the discovery tool summon in ; the usage statistics revealed an increased use of full-text downloads and link resolver software but a decrease in the use of core subject databases. the usage statistics showed promising results, but way recommended further studies of usage statistics over a longer period of time to better understand how discovery tools affect entire library collections. north carolina state university libraries released a final report about their usability study of summon. the results of these usability studies were similar to other studies of discovery tools: users were satisfied with the ability to search the library catalog and article databases with a single search, but users had mixed results with known-item searching and confusion about narrowing facets and results ranking. although several additional academic libraries have conducted usability studies of encore, summon, and ebsco discovery service, the results have not yet been published. information technology and libraries | march only one usability study of ebsco discovery service was found. in a study with six participants, williams and foster found users were satisfied and able to adapt to the new system quickly but did not take full advantage of the rich feature set. combined with the rapid changes in these tools, the literature illustrates a current need for more usability studies related to discovery tools. the necessary focus on specific software implementations and different study designs make it difficult to identify common themes. additional usability studies will offer greater breadth and depth to the current dialogue about discovery tools. this article will help fill the gap by presenting results from a usability study of ebsco discovery service. publishing such usability results of discovery tools will inform institutional decisions, improve user experiences, and advance the tools’ content, features, and interface design. in addition, libraries will be able to more thoroughly modernize library catalogs to meet users’ changing needs and expectations as well as keep pace with the evolution of the web. method james madison university libraries’ usability lab features one workstation with two pieces of usability software: techsmith’s morae (version ) (http://www.techsmith.com/morae.asp), which records screen captures of participant actions during the usability studies, and the usability testing environment (ute) (version ), which presents participants with tasks in a web-browser environment. the ute also presents end-of-task questions to measure time on task and task success. the study of eds, conducted in october , was covered by an institutional review board – approved protocol. participants were recruited for the study through a bulk email sent to all students and faculty. interested respondents were randomly selected to include a variety of grade levels and majors for students and years of service and disciplines taught for faculty members. the study included ten participants with ranging levels of experience: two freshman, two sophomores, two juniors, one senior, one graduate student, and two faculty members. three of the participants were from the school of business, one from education, two from the arts and humanities, and two from the sciences. the remaining two participants had dual majors in the humanities and the sciences. a usability rule of thumb is that at least five users will reveal more than percent of usability issues. because the goal was to observe a wide range of user behaviors and usability issues, and to gather data about satisfaction from a variety of perspectives, this study used two users of each grade level plus two faculty participants (for a total of ten) to provide as much heterogeneity as possible. student participants were presented with ten pre–study questions, and faculty participants were asked nine pre–study questions (see appendix a). the pre–study questions were intended to http://www.techsmith.com/morae.asp usability test results for a discovery tool in an academic library | fagan et al gather information about participants’ background, including their time at jmu, their academic discipline, and their experience with the library website, the ebscohost interface, the library catalog, and library instruction. since participants were anonymous, we hoped their answers would help us interpret unusual comments or findings. pre–test results were not used to form comparison groups (e.g., freshmen versus senior) because these groups would not be representative of their larger populations. these questions were followed by a practice task to help familiarize participants with the testing software. the study consisted of nine tasks designed to showcase usability issues, show the researchers how users behaved in the system, and measure user satisfaction. appendix b lists the tasks and what they were intended to measure. in designing the test, determining success on some tasks seemed very objective (find a video about a given topic) while others appeared to be more subjective (those involving relevance judgments). for this reason, we asked participants to provide satisfaction information on some tasks and not others. in retrospect, for consistency of interpretation, we probably should have asked participants to rate or comment on every task. all of the tasks were presented in the same order. tasks were completed either by clicking “answer” and answering a question (multiple choice or typed response), or by clicking “finished” after navigating to a particular webpage. participants also had the option to skip the task they were working on and move to the next task. allowing participants to skip a task helps differentiate between genuinely incorrect answers and incorrect answers due to participant frustration or guessing. a time limit of minutes was set for tasks – , while tasks and were given time limits of minutes, after which the participant was timed out. time limits were used to ensure participants were able to complete all tasks within the agreed-upon session. average time on task across all tasks was minute, seconds. after the study was completed, participants were presented with the system usability scale (sus), a ten-item scale using statements of subjective assessment and covering a variety of aspects of system usability. sus scores, which provide a numerical score out of , are affected by the complexity of both the system and the tasks users may have performed before taking the sus. the sus was followed by a post–test consisting of six open-ended questions, plus one additional question for faculty participants, intended to gather more qualitative feedback about user satisfaction with the system (see appendix a). a technical glitch with the ute software affected the study in two ways. first, on seven of the ninety tasks, the ute failed to enforce the five-minute maximum time limit, and participants exceeding a task’s time limit were allowed to continue the task until they completed or skipped the task. one participant exceeded the time limit on task while three of these errors occurred during both tasks and . this problem potentially limits the ability to compare the average time on task across tasks; however, since this study used time on task in a descriptive rather than comparative way, the impact on interpreting results is minimal. the seven instances in which the glitch occurred were included in the average time on task data found in figure because the times information technology and libraries | march were not extreme and the time limit had been imposed mostly to be sure participants had time to complete all the tasks. a second problem with the ute was that it randomly and prematurely aborted some users’ tasks; when this happened, participants were informed that their time had run out and were then moved on to the next task. this problem is more serious because it is unknown how much more time or effort the participant would have spent on the task or whether they would have been more successful. because of this, the results below specify how many participants were affected for each task. although this was unfortunate, the results of the participants who did not experience this problem still provide useful cases of user behavior, especially because this study does not attempt to generalize observed behavior or usability issues to the larger population. although a participant mentioned a few technical glitches during testing to the facilitator, the extent of software errors was not discovered until after the tests were complete (and the semester was over) because the facilitator did not directly observe participants during sessions. results the participants were asked several pre–test questions to learn about their research habits. all but one participant indicated they used the library website no more than six times per month (see figure ). common tasks this study’s student participants said they performed on the website were searching for books and articles, searching for music scores, “research using databases,” and checking library hours. the two faculty participants mentioned book and database searches, electronic journal access, and interlibrary loan. participants were shown the quick search widget and were asked “how much of the library’s resources do you think the quick search will search?” seven participants said “most”; only one person, a faculty member, said it would search “all” the library’s resources. figure . monthly visits to library website < visit ( ) - visits ( ) - visits ( ) > visits ( ) usability test results for a discovery tool in an academic library | fagan et al when shown screenshots of the library catalog and an ebscohost database, seven participants were sure they had used leo library catalog, and three were not sure. three indicated that they had used an ebsco database before, five had not, and two were not su re. participants were also asked how often they had used library resources for assignments in their major field of study; four said “often,” two said “sometimes,” one “rarely/never,” and one “very often.” students were also asked “has a librarian spoken to a class you’ve attended about library research?” and two said yes, five said no, and one was not sure. a “practice task” was administered to ensure participants were comfortable with the workstation and software: “use quick search to search a topic relating to your major/discipline or another topic of interest to you. if you were writing a paper on this topic how satisfied would you be with these results?” no one selected “no opinion” or very unsatisfied”; sixty percent were “very satisfied” or “satisfied” with their results; forty percent were “somewhat unsatisfied.” figure shows the time spent on each task, while figure describes participants’ success on the tasks. task task task task task task task task task no. of responses (not including timeouts) avg. time on task (in seconds) * * * standard deviation *includes time(s) in excess of the set time limit. excess time allowed by software error. figure . average time spent on tasks task task task task task task task task task t im e o n t a sk ( in s e co n d s) average time for all tasks (not including timeouts) information technology and libraries | march the first task (“what was the last thing you searched for when doing a research assignment for class? use quick search to re-search for this.”) started participants on the library homepage. participants were then asked to “tell us how this compared to your previous experience” using a text box. the average time on task was almost minutes; however one faculty participant took more than minutes on this task; if his or her time was removed, the time on task average was minute, seconds. figure shows the participants’ search terms and their comments. task task task task task task task task task how success deter- mined users only asked to provide feedback valid typed-in response provided how many subtasks completed (out of ) how many subtasks completed (out of ) correct multiple choice answer how many subtasks completed (out of ) end task at correct web location how many subtasks complete d (out of ) how many subtasks completed (out of ) p n/a correct timeout correct * ** p n/a correct * correct correct ** p n/a correct * incorrect correct p n/a correct * correct skip p n/a correct* correct correct p n/a correct * correct correct ** p n/a correct * correct correct p n/a correct * correct skip timeout ** p n/a correct * skip correct correct p n/a correct * correct skip note: “timeout” indicates an immediate timeout error. users were unable to take any action on the task. *user experienced a timeout error while working on the task. this may have affected their ability to complete the task. **user did not follow directions. figure . participants’ success on tasks usability test results for a discovery tool in an academic library | fagan et al participant jmu status major/discipline search terms p faculty geology large low shear wave velocity province comments: ebsco did a fairly complete job. there were some irrelevant results that i don’t remember seeing when i used georef. p faculty computer information systems & management science (statistics) student cheating comments: this is a topic that i am somewhat familiar with the related literature. i was pleased with the diversity of journals that were found in the search. the topics of the articles was right on target. the recency of the articles was great. this is a topic for which i am somewhat familiar with the related literature. i was impressed with the search results regarding: diversity of journals; recency of articles; just the topic in articles i was looking for. p graduate student education death of a salesman comments: there is a lot of variety in the types of sources that quick search is pulling up now. i would still have liked to see more critical sources on the play but i could probably have found more results of that nature with a better search term, such as “death of a salesman criticism.” p st year voice performance current issues in russia comments: it was somewhat helpful in the way that it gave me information about what had happened in the past couple months, but not what was happening now in russia. p rd year nursing uninsured and health care reform comments: the quick search gave very detailed articles i thought, which could be good, but were not exactly what i was looking for. then again, i didn’t read all these articles either p st year history headscarf law comments: this search yielded more results related to my topic. i needed other sources for an argument on the french creating law banning religious dress and symbols in school. using other methods with the same keyword, i had an enormous amount of trouble finding articles that pertained to my essay. p rd year english jung comments: i like the fact that it can be so defined to help me get exactly what i need. p th year spanish restaurant industry comments: this is about the same as the last time that i researched this topic. p nd year hospitality aphasia comments: there are many good sources, however there are also completely irrelevant sources. p nd year management rogers five types of feedback comments: there is not many documents on the topic i searched for. this may be because the topic is not popular or my search is not specific/too specific. figure . participants’ search terms and comments information technology and libraries | march the second task started on the library homepage and asked participants to find a video related to early childhood cognitive development. this task was chosen because jmu libraries have significant video collections and because the research team hypothesized users might have trouble because there was no explicit way to limit to videos at the time. the average time on this task was two minutes, with one person experiencing an arbitrary time out by the software. participants were judged to be successful on this task by the researchers if they found any video related to the topic. all participants were successful on this task, but four entered, then left the discovery tool interface to complete the task. five participants looked for a video search option in the drop-down menu, and of these, three immediately used something other than quick search when they saw that there was no video search option. of those who tried quick search, six opened the source type facet in eds search results and four selected a source type limit, but only two selected a source type that led directly to success (“non-print resources”). task started participants in eds (see figure ) and asked them to search on speech pathology, find a way to limit search results to audiology, and limit their search results to peer-reviewed sources. participants spent an average of minute, seconds on this task, with five participants being artificially timed out by the software. participants’ success on this task was determined by the researchers’ examination of the number of subtasks they completed. the three subtasks consisted of successfully searching for the given topic (speech language pathology) limiting the search results to audiology, and further limiting the results to peer reviewed sources. four participants were able to complete all three subtasks, including two who were timed out. (the times for those who were timed out were not included in time on task averages, but they were given credit for success.) five completed just two of the subtasks, failing to limit to peerreviewed; one of these because of a timeout. it was unclear why the remaining participants did not attempt to alter the search results to “peer reviewed.” looking at the performed actions, six of the ten typed “and audiology” into search keywords to narrow the search results, while one found and used “audiology” in the subject facet on the search results page. six participants found and used the “scholarly (peer reviewed) journals” checkbox limiter. usability test results for a discovery tool in an academic library | fagan et al figure . ebsco discovery service interface beginning with the results they had from task , task asked participants to find more recent sources and to select the most recent source available. task success was measured by correct completion of two subtasks: limiting the search results to the last five years and finding the most recent source. the average time on task was minute, seconds, with three artificial timeouts. of those who did not time out, all seven were able to limit their sources to be more recent in some way, but only three were able to select the most recent source. in addition to this being a common research task, the team was interested to see how users accomplished this task. three typed in the limiter in the left-hand column, two typed in the limiter on the advanced search screen, and two used the date slider. two participants used the “sort” drop-down menu to change the sort order to “date descending,” which helped them complete this task. other participants changed the dates, and then selected the first result, which was not the most recent. task , which started within eds, asked participants to find a way to ask a jmu librarian for help. the success of this task was measured by whether they reached the correct url for the ask-a- information technology and libraries | march librarian page; eight of the ten participants were successful. this task took an average of only seconds to complete, and eight of the ten used the ask-a-librarian link at the top of the page. of the two unsuccessful participants, one was timed out, while another clicked “search modes” for no apparent reason, then clicked back and decided to finish the task. task started in the eds interface and asked participants to locate the journal yachting and boating world and select the correct coverage dates and online status from a list of four options; participants were deemed successful at two subtasks if they selected the correct option and successful at one subtask if they chose an option that was partially correct. participants took an average of two minutes on this task; only five answered correctly. during this task, three participants used the ebsco search option “so journal title/source,” four used quotation marks, and four searched or re-searched with the “title” drop-down menu option. three chose the correct dates of coverage, but were unable to correctly identify the online availability. it is important to note that only searching and locating the journal title were accomplished with the discovery tool; to see dates of coverage and online availability, users clicked jmu’s link resolver button, and the resulting screen was served from serials solutions’ article linker product. although some users spent more time than perhaps was necessary using the eds search options to locate the journal, the real barriers to this task were encountered when trying to interpret the serials solutions screen. task , where participants started in eds, was designed to determine whether users could navigate to a research database outside of eds. users were asked to look up the sculpture genius of mirth and were told the library database camio would be the best place to search. they were instructed to “locate this database and find the sculpture.” the researcher observed the recordings to determine success on this task, which was defined as using camio to find the sculpture. participants took an average of minute, seconds on this task; seven were observed to complete the task successfully, while three chose to skip the task. to accomplish this task, seven participants used the jmu research databases link in the header navigation at some point, but only four began the task by doing this. six participants began by searching within eds. the final two tasks started on the library homepage and were a pair: participants were asked to find two books and two recent, peer-reviewed articles (from the last five years) on rheumatoid arthritis. task asked them to use the library’s eds widget, quick search, to accomplish this, and task asked them to accomplish the same task without using quick search. when they found sources, they were asked to enter the four relevant titles in a text-entry box. the average time spent on these tasks was similar: about four minutes per task. comparing these tasks was somewhat confusing because some participants did not follow instructions. user s uccess was determined by the researchers’ observation of how many of the four subtasks the user was able to complete successfully: find two books, find two articles, limit to peer reviewed, and select articles from last five years (with or without using a limiter); figure shows their success. usability test results for a discovery tool in an academic library | fagan et al looking at the seven users who used quick search on the quick search tasks, six limited to “scholarly (peer reviewed) journals”; six limited to the last five years; and seven narrowed results using the source type facet. the average number of subtasks completed on task eight was . out of . looking at the seven users who followed instructions and did not use quick search on task , all began with the library catalog and tried to locate articles within the library catalog. the average number of subtasks completed on task was . out of . some users tried to locate articles by setting the catalog’s material type drop-down menu to “periodicals” and others used the catalog’s “periodical” tab, which performed a title keyword search of the e-journal portal. for task , only two users eventually chose a research database to find articles. user behavior can only be compared for the six users (all students) who followed instructions on both tasks; a summary is provided in figure . after completing all nine tasks, participants were presented with the system usability scale. eds scored out of . following the sus, participants were asked a series of post–test questions. only one of the faculty members chose to answer the post–test questions. when asked how they would use quick search, all eight students explicitly mentioned class assignments, and the participating faculty member replied “to search for books.” two students mentioned books specifically, while the rest used the more generic term “sources” to describe items for which they would search. when asked “when would you not use this search tool?” the faculty member said “i would just have to get used to using it. i mainly go to [the library catalog] and then research databases.” responses from the six students who answered this question were vague and hard to categorize: • “not really sure for more general question/learning” • “when just browsing” • “for quick answers” • “if i could look up the information on the internet” • “when the material i need is broad” • “basic searching when you do not need to say where you got the info from” when asked for the advantages of quick search, four specifically mentioned the ability to narrow results, three respondents mentioned “speed,” three mentioned ease of use, and three mentioned relevance in some way (e.g., “it does a pretty good job associating keywords with sources”). two mentioned the broad coverage and one compared it to google, “which is what students are looking for.” when asked to list disadvantages, the faculty member mentioned he/she was not sure what part of the library home page was actually “quick search,” and was not sure how to get to his/her library account. three students talked about quick search being “overwhelming” or “confusing” because of the many features, although one of these also stated, “like anything you need to learn in order to use it efficiently.” one student mentioned the lack of an audio recording limit and another said “when the search results come up it is hard to tell if they are usable results.” information technology and libraries | march knowing that quick search may not always provide the best results, the research team also asked users what they would do if they were unable to find an item using quick search. a faculty participant said he or she would log into the library catalog and start from there. five students mentioned consulting a library staff member in some fashion. three mentioned moving on from library resources, although not necessarily as their first step. one said “find out more information on it to help narrow down my search.” only one student mentioned the library catalog or any other specific library resource. when participants were asked if “quick search” was an appropriate name, seven agreed that it was. of those who did not agree, one participant’s comment was “not really, though i don’t think it matters.” and another’s was “i think it represents the idea of the search, but not the action. it could be quicker.” the only alternative name suggestion was “search tool.” web traffic analysis web traffic through quick search and in eds provides additional context for this study’s results. during august–december , quick search was searched , times from the library homepage. this is an increase from traffic into the previous widget in this location that searched the catalog, which received , searches during the same period in . even adjusting for an approximately percent increase in website traffic from to , this is an increase of percent. interestingly, the traffic to the most popular link on the library homepage, research databases, went from , in to , in , a decrease of percent when adjusting for the change in website traffic. during fall , percent of quick search searches from the homepage were executed using at least one drop-down menu. twelve percent changed quick search’s first drop-down menu to something other than the keyword default, with “title” being the most popular option ( percent of searches) followed by author ( percent of searches). twenty percent of users changed the second drop-down option; “just articles” and “just books” were the most popular options, garnering percent and percent of searches, respectively, followed by “just scholarly articles,” which accounted for percent of searches. looking at ebsco’s statistical reports for jmu’s implementation of eds, there were , sessions and approximately , searches during august–december . this means about percent of eds sessions were launched using quick search from the homepage. there were an average of . searches per session, which is comparable to past behavior in jmu’s other ebscohost databases. discussion usability test results for a discovery tool in an academic library | fagan et al the goal of this study was to gather initial data about user behavior, usability issues, and user satisfaction with discovery tools. the task design and technical limitations of the study mean that comparing time on task between participants or tasks would not be particularly illuminating; and, while the success rates on tasks are interesting, they are not generalizable to the larger jmu population. instead, this study provided observations of user behavior that librarians can use to improve services, it suggested some “quick fixes” to usability issues, and it pointed to several research questions. when possible, these observations are supplemented by comparisons between this study and the only other published usability study of eds. this study confirmed a previous finding of user studies of federated search software and discovery tools: students have trouble determining what is searched by various systems. on the tasks in which they were asked to not use quick search to find articles, participants tried to search for articles in the library catalog. although all but one of this study’s participants correctly answered that quick search did not search “all” library resources, seven thought it searched “most.” both “most” or “some” would be considered correct; however, it is interesting that answering this question more specifically is challenging even for librarians. many journals in subject article indexes and abstracts are included in the eds foundation index; furthermore, jmu’s implementation of eds includes all of jmu’s ebsco subscription resources as well, making it impractical to assemble a master list of indexed titles. of course, there are numerous online resources with contents which may never be included in a discovery tool, such as political voting records, ethnographic files, and financial data. users often have access to these resources through their library. however, if they do not know the library has a database of financial data, they will certainly not consider this content in their response to a question of how many of the library resources are included in the discovery tool. as discovery tools begin to fulfill users’ expectations for a “single search,” libraries will need to share best practices for showcasing valuable, useful collections that fall outside the discovery tool’s scope or abilities. this is especially critical when reviewing the percent increase in homepage traffic to the homepage search widget compared with the percent decrease in homepage traffic to the research databases page. it is important to note these trends do not mean the library’s other research databases have fallen in usage by percent. though there was not a comprehensive examination of usage statistics, spot-checking suggested ebsco and non-ebsco subject databases had both increases and decreases in usage from previous years. another issue libraries should consider, especially when preparing for instruction classes, is that users do not seem to understand which information needs are suited to a discovery tool versus the catalog or subject-specific databases. several tasks provided additional information about users’ mental models of the tool, which may help libraries make better decisions about navigation customizations in discovery tool interfaces and on library websites. task was designed to discover whether users could find their way to a database outside of eds if they knew they needed to use a specific database. six participants, including one of the faculty members, began by searching eds for the name of the sculpture and/or the database name. on task , a graduate information technology and libraries | march student who searched on “death of a salesman” and was asked to comment on how quick search results compared to his or her previous experience, said, “i would still have liked to see more critical sources on the play but i could probably have found more results of that nature with a better search term, such as ‘death of a salesman criticism.’” while true, most librarians would suggest using a literary criticism database, which would target this information need. librarians may have differing opinions regarding the best research starting point, but their rationale would be much different than that of the students in this study. this study’s participants said they would use quick search/eds when they were doing class work or research, but would not use it for general inquiries. if librarians were to list which user information needs are best met by a discovery tool versus a subject-specific database, the types of information needs listed would be much more numerous and diverse, regardless of differences over how to classify them. in addition to helping users choose between a discovery tool or a subject-specific database, libraries will need to conceptualize how users will move in and out of the discovery tool to other library resources, services, and user accounts. while users had no trouble finding the ask-a- librarian link in the header, it might have been more informative if users started from a search- results page to see if they would find the right-hand column’s ask-a-librarian link or links to library subject guides and database lists. discovery tools vary in their abilities to connect users with their online library accounts and are changing quickly in this area. this study also provided some interesting observations about discovery tool interfaces. the default setting for ebsco discovery service is a single search box. however, this study suggests that while users desire a single search, they are willing to use multiple interface options. this was supported by log analysis of the library’s locally developed entry widget, quick search, in which percent of searches included the use of a drop-down menu. on the first usability task, users left quick search’s options set to the default. on other tasks, participants frequently used the drop- down menus and limiters in both quick search and eds. for example, on task , which asked them to look for videos, five users looked in the quick search format drop-down menu. on the same task within eds, six users attempted to use the source type facet. use of limiters was similarly observed by williams and foster in their eds usability study. one eds interface option that was not obvious to participants was the link to change the sort order. when asked to find the most recent article, only two participants changed the sort option. most others used the date input boxes to limit their search, then selected the first result even thought it was not the most recent one. it is unclear whether the participant assumed the first result was the most recent or whether they could not figure out how to display the most recent sources. finding a journal title from library homepages has long been a difficult task, and this study provided no exception, even with the addition of a discovery tool. it is important to note that the standard eds implementation would include a “publications” or “journals a–z” link in the header; usability test results for a discovery tool in an academic library | fagan et al in eds, libraries can customize the text of this link. jmu did not have this type of link enabled in our test, since the hope was that users could find journal titles within the eds results. however, neither eds nor the quick search widget’s search interfaces offered a way to limit the search to a journal title at the time of this study. during the usability test, four participants changed the field search drop-down menu to “title” in eds, and three participants changed the eds field search drop-down menu to “so journal title/source,” which limits the search to articles within that journal title. while both of these ideas were good, neither one resulted in a precise results set in eds for this task unless the user also limited to “jmu catalog only,” a nonintuitive option. since the test, jmu has added a “journal titles” option to quick search that launches the user’s search into the journal a–z list (provided by serials solutions). in two months after the change (february and march ), only searches were performed with this option. this was less than percent of all searches, indicating that while it may be an important task, it is not a popular one. like many libraries with discovery tools, jmu added federated search capabilities to eds using ebscohost integrated search software in an attempt to draw some traffic to databases not included in eds (or not subscribed to through ebsco by jmu), such as mla international bibliography, scopus, and credo reference. links to these databases appeared in the upper-right- hand column of eds during the usability study (see figure .) usage data from ebsco showed that less than percent of all jmu’s eds sessions for fall included any interaction with this area. likewise, williams and foster observed their participants did not use their federated search until explicitly asked to do so. perhaps users faced with discovery tool results simply have no motivation to click on additional database results. since the usability test, jmu has replaced the right-hand column with static links to ask-a-librarian, subject guides, and research database lists. readers may wonder why one of the most common tasks, finding a specific book title, was not included in this usability study; this was because jmu libraries posed this task in a concurrent homepage usability study. on that study, twenty of the twenty-five participants used quick search to find the title “pigs in heaven” and choose the correct call number. eleven of the twenty used the quick search drop-down menu to choose a title search option, further confirming users’ willingness to limit up-front. the average time on this task was just under a minute, and all participants completed this task successfully, so this task was not repeated in the eds usability test. other studies have reported trouble with this type of task; much could depend on the item chosen as well as the tool’s relevance ranking. user satisfaction with eds can be summarized from the open-ended post–study questions, from the responses to task (figure ), and the sus scale. answers to the post–study questions indicated participants liked the ability to narrow results, the speed and ease of use, and relevance of the system. a few participants did describe the system as being “overwhelming” or “confusing” because of the many features, which was also supported by the sus scores. jmu has been using the sus to understand the relative usability of library systems. the sus offers a benchmark for system improvement; for example, ebsco discovery service received an sus of only in spring (n information technology and libraries | march = ) but a on this study in fall (n = ). this suggests the interface has become more usable. in , jmu libraries also used the sus to test the library catalog’s classic interface as well as a vufind interface to the library catalog, which received scores of (n = ) and (n = ), respectively. the differences between the catalog scores and eds indicate an important distinction between usability and usefulness, with the latter concept encompassing a system’s content and capabilities. the library catalog is, perhaps, a more straightforward tool than a discovery tool and attempts to provide access to a smaller set of information. it has none of the complexity involved in finding article-level or book chapter information. all else being equal, simpler tools will be more usable. in an experimental study, tsakonas and paptheodorou found that while users did not distinguish between the concepts of usability and usefulness, they prefer attributes composing a useful system in contrast to those supporting usability. discovery tools, which support more tasks, must make compromises in usability that simpler systems can avoid. in their study of eds, williams and foster also found overall user satisfaction with eds. their participants made positive comments about the interface as well as the usefulness and relevance of the results. jmu passed on several suggestions to ebsco related to eds based on the test results. ebsco subsequently added “audio” and “video” to the source types, which enabled jmu to add a “just videos at jmu” option to quick search. while it is confusing that “audio” and “video” source types currently behave differently than the others in eds, in that they limit to jmu’s catalog as well as to the source type, this behavior produces what most local users expect. a previous usability study of worldcat local showed users have trouble discriminating between source types in results lists, so the source types facet is important. another piece of feedback provided to ebsco was that on the task where users needed to choose the most recent result, only two of our participants sorted by date descending. perhaps the textual appearance of the sort option (instead of a drop-down menu) was not obvious to participants (see figure ); however, williams and foster did not observe this to be an issue in their study. future research the findings of this study suggest many avenues for future research. libraries will need to revisit the scope of their catalogs and other systems to keep up with users’ mental models and information needs. catalogs and subject-specific databases still perform some tasks much better than discovery tools, but libraries will need to investigate how to situate the discovery tool and specialized tools within their web presence in a way that will make sense to users. when should a user be directed to the catalog versus a discovery tool? what items should libraries continue to include in their catalogs? what role do institutional repositories play in the suite of library tools, and how does the discovery tool connect to them (or include them?) how do library websites begin to make sense of the current state of library search systems? above all, are users able to find the best resources for their research needs? although research on searchers’ mental models has been extensive, librarians’ mental models have not been studied as such. yet placing the usability test results for a discovery tool in an academic library | fagan et al discovery tool among the library’s suite of services will involve compromises between these two models. another area needing research is how to instruct users to work with the large numbers of results returned by discovery tools. in subject-specific databases, librarians often help users measure the success of their strategy—or even their topic—by the number of results returned: in criminal justice abstracts, , results means a topic is too broad or the search strategy needs refinement. in a discovery tool, a result set this large will likely have some good results on the first couple of pages if sorted by relevance; however, users will still need to know how to grow or reduce their results sets. participants in this study showed a willingness to use limiters and other interface features, but not always the most helpful ones. when asked to narrow a broad subject on task of this study, only one participant chose to use the “subject” facet even when the subtopic, audiology, was clearly available. most added search terms. it will be important for future studies to investigate the best way for users to narrow large results set in a discovery tool. this study also suggested possible areas of investigation for future user studies. one interesting finding related to this study’s users’ information contexts was that when users were asked to search on their last research topic, it did not always match up with their major: a voice performance student searched on “current issues in russia,” and the hospitality major searched on “aphasia.” to what extent does a discovery tool help or hinder students who are searching outside their major area of study? one of jmu’s reference librarians noted that while he would usually teach a student majoring in a subject how to use that subject’s specific indexes, as opposed to a discovery tool, a student outside the major might not need to learn the subject-specific indexes for that subject and could be well served by the discovery tool. future studies could also investigate the usage and usability of discovery tool features in order to continue informing library customizations and advice to vendors. for example, this study did not have a task related to logging into a patron account or requesting items, but that would be good to investigate in a follow-up study. another area ripe for further investigation is discovery tool limiters. this study’s participants frequently attempted to use limiters, but didn’t always choose the correct ones for the task. what are the ideal design choices for making limiters intuitive? this study found almost no use of the embedded federated search add-on: is this true at other institutions? finally, this study and others reveal difficulty in distinguishing source types. development and testing of interface enhancements to support this ability would be helpful to many libraries’ systems. conclusion this usability test of a discovery tool at james madison university did not reveal as many interface-specific findings as it did questions about the role of discovery tools in libraries. users were generally able to navigate through the quick search and eds interfaces and complete tasks successfully. tasks that are challenging in other interfaces, such as locating journal articles and discriminating between source types, continued to be challenging in a discovery tool interface. information technology and libraries | march this usability test suggested that while some interface features were heavily used, such as drop - down limits and facets, other features were not used, such as federated search results. as discovery tools continue to grow and evolve, libraries should continue to conduct usability tests, both to find usability issues and to understand user behavior and satisfaction. although discovery tools challenge libraries to think not only about access but also about the best research pathways for users, they provide users with a search that more closely matches their expectations. acknowledgement the authors would like to thank patrick ragland for his editorial assistance in preparing this manuscript. correction april , : at the request of the author, this article was revised to remove a link to a website. references . emily alling and rachael naismith, “protocol analysis of a federated search tool: designing for users,” internet reference services quarterly , no. ( ): , http://scholarworks.umass.edu/librarian_pubs/ / (accessed jan. , ); frank cervone, “what we've learned from doing usability testing on openurl resolvers and federated search engines,” computers in libraries , no. ( ): ; sara randall, “federated searching and usability testing: building the perfect beast,” serials review , no. ( ): – , doi: . /j.serrev. . . ; ed tallent, “metasearching in boston college libraries —a case study of user reactions,” new library world , no. ( ): - , doi: . / . . s. c. williams and a. k. foster, “promise fulfilled? an ebsco discovery service usability study,” journal of web librarianship , no. ( ), http://www.tandfonline.com/doi/pdf/ . / . . (accessed jan. , ). . janet k. chisman, karen r. diller, and sharon l. walbridge, “usability testing: a case study,” college & research libraries , no. (november ): – , http://crl.acrl.org/content/ / / .short (accessed jan. , ); frances c. johnson and jenny craven, “beyond usability: the study of functionality of the . online catalogue,” new review of academic librarianship , no. ( ): – , doi: . / (accessed jan, , ); jennifer e. knievel, jina choi wakimoto, and sara holladay, “does interface design influence catalog use? a case study,” college & research libraries , no. (september ): – , http://crl.acrl.org/content/ / / .short (accessed jan. , ); jia mi and cathy weng, “revitalizing the library opac: interface, searching, and display challenges,” information technology & libraries , no. (march ): – , http:// - http://scholarworks.umass.edu/librarian_pubs/ / http://www.tandfonline.com/doi/pdf/ . / . . http://crl.acrl.org/content/ / / .short http://crl.acrl.org/content/ / / .short http:// -www.ala.org.sapl.sat.lib.tx.us/ala/mgrps/divs/lita/publications/ital/ / /mi.pdf usability test results for a discovery tool in an academic library | fagan et al www.ala.org.sapl.sat.lib.tx.us/ala/mgrps/divs/lita/publications/ital/ / /mi.pdf (accessed jan. , ). . karen calhoun, “the changing nature of the catalog and its integration with other discovery tools,” http://www.loc.gov/catdir/calhoun-report-final.pdf (accessed mar. , ). . dee ann allison, “information portals: the next generation catalog,” journal of web librarianship , no. ( ): – , http://digitalcommons.unl.edu/cgi/viewcontent.cgi?article= &context=libraryscience (accessed january , ); marshall breeding, “the state of the art in library discovery,” computers in libraries , no. ( ): – ; c. p diedrichs, “discovery and delivery: making it work for users . . . taking the sting out of serials!” (lecture, north american serials interest group, inc. rd annual conference, phoenix, arizona, june – , ), doi: . / ; ian hargraves, “controversies of information discovery,” knowledge, technology & policy , no. (summer ): , http://www.springerlink.com/content/au jr /fulltext.html (accessed jan. , ); jane hutton, “academic libraries as digital gateways: linking students to the burgeoning wealth of open online collections,” journal of library administration , no. ( ): – , doi: . / ; oclc, “online catalogs: what users and librarians want: an oclc report,” http://www.oclc.org/reports/onlinecatalogs/default.htm (accessed mar. ). . c. j. belliston, jared l. howland, and brian c. roberts, “undergraduate use of federated searching: a survey of preferences and perceptions of value-added functionality,” college & research libraries , no. (november ): – , http://crl.acrl.org/content/ / / .full.pdf+html (accessed jan. , ); judith z. emde, sara e. morris, and monica claassen‐wilson, “testing an academic library website for usability with faculty and graduate students,” evidence based library & information practice , no. ( ): – , http://kuscholarworks.ku.edu/dspace/bitstream/ / / /emdee_morris_cw.pdf (accessed jan. , ); karla saari kitalong, athena hoeppner, and meg scharf, “making sense of an academic library web site: toward a more usable interface for university researchers,” journal of web librarianship , no. / ( ): – , http://www.tandfonline.com/doi/abs/ . / (accessed jan. , ); ed tallent, “metasearching in boston college libraries—a case study of user reactions,” new library world , no. ( ): – , doi: . / ; rong tang, ingrid hsieh-yee, and shanyun zhang, “user perceptions of metalib combined search: an investigation of how users make sense of federated searching,” internet reference services quarterly , no. ( ): – , http://www.tandfonline.com/doi/abs/ . /j v n _ (accessed jan. , ). . jody condit fagan, “usability studies of faceted browsing: a literature review,” information technology & libraries , no. ( ): – , http:// -www.ala.org.sapl.sat.lib.tx.us/ala/mgrps/divs/lita/publications/ital/ / /mi.pdf http://www.loc.gov/catdir/calhoun-report-final.pdf http://digitalcommons.unl.edu/cgi/viewcontent.cgi?article= &context=libraryscience http://www.springerlink.com/content/au jr /fulltext.html http://www.oclc.org/reports/onlinecatalogs/default.htm http://crl.acrl.org/content/ / / .full.pdf+html http://kuscholarworks.ku.edu/dspace/bitstream/ / / /emdee_morris_cw.pdf http://www.tandfonline.com/doi/abs/ . / http://www.tandfonline.com/doi/abs/ . /j v n _ information technology and libraries | march http://web .ala.org/ala/mgrps/divs/lita/publications/ital/ / /fagan.pdf (accessed jan. , ). . birong ho, keith kelley, and scott garrison, “implementing vufind as an alternative to voyager’s web voyage interface: one library’s experience,” library hi tech , no. ( ): , doi: . / (accessed jan. , ). . tamar sadeh, “user experience in the library: a case study,” new library world , no. ( ): – , doi: . / (accessed jan. , ). . tod a. olson, “utility of a faceted catalog for scholarly research,” library hi tech , no. ( ): – , doi: . / (accessed jan. , ). . allison, “information portals,” – . . marshall breeding, “plotting a new course for metasearch,” computers in libraries , no. ( ): . . ibid. . dennis brunning and george machovec, “interview about summon with jane burke, vice president of serials solutions,” charleston advisor , no. ( ): – ; dennis brunning and george machovec, “an interview with sam brooks and michael gorrell on the ebscohost integrated search and ebsco discovery service,” charleston advisor , no. ( ): – , http://www.ebscohost.com/uploads/discovery/pdfs/topicfile- .pdf (accessed jan. , ). . ronda rowe, “web-scale discovery: a review of summon, ebsco discovery service, and worldcat local,” charleston advisor , no. ( ): – ; k. stevenson et al., “next-generation library catalogues: reviews of encore, primo, summon and summa,” serials , no. ( ): – . . jason vaughan, “chapter : questions to consider,” library technology reports , no. ( ): ; paula l. webb and muriel d. nero, “opacs in the clouds,” computers in libraries , no. ( ): . . jason vaughan, “investigations into library web scale discovery services,” articles (libraries), paper ( ), http://digitalcommons.library.unlv.edu/lib_articles/ . . marshall breeding, “the state of the art in library discovery,” – ; sharon q. yang and kurt wagner, “evaluating and comparing discovery tools: how close are we towards next generation catalog?” library hi tech , no. ( ): – . . allison, “information portals,” – . . breeding, “the state of the art in library discovery,” – . . galina letnikova, “usability testing of academic library websites: a selective bibliography,” internet reference services quarterly , no. ( ): – . http://web .ala.org/ala/mgrps/divs/lita/publications/ital/ / /fagan.pdf http://www.ebscohost.com/uploads/discovery/pdfs/topicfile- .pdf http://digitalcommons.library.unlv.edu/lib_articles/ usability test results for a discovery tool in an academic library | fagan et al . jeffrey rubin and dana chisnell, handbook of usability testing: how to plan, design, and conduct effective tests, nd ed. (indianapolis, in: wiley, ); joseph s. dumas and janice redish, a practical guide to usability testing, rev. ed. (portland, or: intellect, ). . nicole campbell, ed., usability assessment of library-related web sites: methods and case studies (chicago: library & information technology association, ); elaina norlin and c. m. winters, usability testing for library web sites: a hands-on guide (chicago: american library association, ). . jennifer l. ward, steve shadle, and pam mofield, “user experience, feedback, and testing,” library technology reports , no. ( ): . . ibid. . michael boock, faye chadwell, and terry reese, “worldcat local task force report to lamp,” http://hdl.handle.net/ / (accessed mar. ). . bob thomas and stefanie buck, “oclc’s worldcat local versus iii’s webpac: which interface is better at supporting common user tasks?” library hi tech , no. ( ): – . . oclc, “some findings from worldcat local usability tests prepared for ala annual,” http://www.oclc.org/worldcatlocal/about/ usf_some_findings_about_worldcat_local.pdf (accessed mar. , ). . ibid., . . doug way, “the impact of web-scale discovery on the use of a library collection,” serials review , no. ( ): . . north carolina state university libraries, “final summon user research report,” http://www.lib.ncsu.edu/userstudies/studies/ _summon/ (accessed mar. , ). . alesia mcmanus, “the discovery sandbox: aleph and encore playing together,” http://www.nercomp.org/data/media/discovery% sandbox% mcmanus.pdf (accessed mar. , ); prweb, “deakin university in australia chooses ebsco discovery service,” http://www.prweb.com/releases/deakin/chooseseds/prweb .htm (accessed mar. , ); university of manitoba, “summon usability: partnering with the vendor,” http://prezi.com/icxawthckyhp/summon-usability-partnering-with-the-vendor (accessed mar. , ). . williams and foster, “promise fulfilled?” . jakob nielsen, “why you only need to test with users,” http://www.useit.com/alertbox/ .html (accessed aug. , ). . john brooke, “sus: a ‘quick and dirty’ usability scale,” in usability evaluation in industry, ed. p. w. jordanet al. (london: taylor & francis, ), http://www.usabilitynet.org/trump/documents/suschapt.doc (accessed apr. , ). . williams and foster, “promise fulfilled?” http://hdl.handle.net/ / http://www.oclc.org/worldcatlocal/about/ usf_some_findings_about_worldcat_local.pdf http://www.lib.ncsu.edu/userstudies/studies/ _summon/ http://www.nercomp.org/data/media/discovery% sandbox% mcmanus.pdf http://www.prweb.com/releases/deakin/chooseseds/prweb .htm http://prezi.com/icxawthckyhp/summon-usability-partnering-with-the-vendor/ http://www.useit.com/alertbox/ .html http://www.usabilitynet.org/trump/documents/suschapt.doc information technology and libraries | march . seikyung jung et al., “libraryfind: system design and usability testing of academic metasearch system,” journal of the american society for information science & technology , no. ( ): – ; williams and foster, “promise fulfilled?”; laura wrubel and kari schmidt, “usability testing of a metasearch interface: a case study,” college & research libraries , no. ( ): – . . williams and foster, “promise fulfilled?” . letnikova, “usability testing of academic library websites,” – ; tom ipri, michael yunkin, and jeanne m. brown, “usability as a method for assessing discovery,” information technology & libraries , no. ( ): – ; susan h. mvungi, karin de jager, and peter g. underwood, “an evaluation of the information architecture of the uct library web site,” south african journal of library & information science , no. ( ): – . . williams and foster, “promise fulfilled?” . ward et al., “user experience, feedback, and testing,” . . giannis tsakonas and christos papatheodorou, “analysing and evaluating usefulness and usability in electronic information services,” journal of information science , no. ( ): – . . williams and foster, “promise fulfilled?” . bob thomas and stefanie buck, “oclc’s worldcat local versus iii’s webpac: which interface is better at supporting common user tasks?” library hi tech , no. ( ): – . . williams and foster, “promise fulfilled?” . tracy gabridge, millicent gaskell, and amy stout, “information seeking through students’ eyes: the mit photo diary study,” college & research libraries , no. ( ): – ; yan zhang, “undergraduate students’ mental models of the web as an information retrieval system,” journal of the american society for information science & technology , no. ( ): – ; brenda reeb and susan gibbons, “students, librarians, and subject guides: improving a poor rate of return,” portal: libraries and the academy , no. ( ): – ; alexandra dimitroff, “mental models theory and search outcome in a bibliographic retrieval system,” library & information science research , no. ( ): – . usability test results for a discovery tool in an academic library | fagan et al appendix a task pre–test : please indicate your jmu status ( st year, nd year, rd year, th year, graduate student, faculty, other) pre–test : please list your major(s) or area of teaching (open ended) pre–test : how often do you use the library website? (less than once a month, – visits per month, – visits per month, more than visits per month) pre–test : what are some of the most common things you currently do on the library website? (open ended) pre–test : how much of the library’s resources do you think the quick search will search? (less than a third, less than half, half, most, all) pre–test : have you used leo? (show screenshot on printout) (yes, no, not sure) pre–test : have you used ebsco? (show screenshot on printout) (yes, no, not sure) pre–test (student participants only): how often have you used library web resources for course assignments in your major? (rarely/never, sometimes, often, very often) pre–test (student participants only): how often have you used library resources for course assignments outside of your major? (rarely/never, sometimes, often, very often) pre–test (student participants only): has a librarian spoken to a class you've attended about library research? (yes, no, not sure) pre–test (faculty participants only): how often do you give assignments that require the use of library resources? (rarely/never, sometimes, often, very often) pre–test (faculty participants only): how often have you had a librarian visit one of your classes to teach your students about library research? (rarely/never, sometimes, often, very often) post–test : when would you use this search tool? post–test : when would you not use this search tool? post–test : what would you say are the major advantages of quick search? information technology and libraries | march post–test : what would you say are the major problems with quick search? post–test : if you were unable to find an item using quick search/ebsco discovery service what would your next steps be? post–test : do you think the name “quick search” is fitting for this search tool? if not, what would you call it? post–test (faculty participants only): if you knew students would use this tool to complete assignments would you alter how you structure assignments and how? appendix b task purpose • practice task: use quick search to search a topic relating to your major / discipline or another topic of interest to you. if you were writing a paper on this topic how satisfied would you be with these results? help users get comfortable with the usability testing software. also, since the first time someone uses a piece of software involves behaviors unique to that case, we wanted participants’ first use of eds to be with a practice task. . what was the last thing you searched for when doing a research assignment for class? use quick search to re-search for this. tell us how this compared to your previous experience. having participants re-search a topic with which they had some experience and interest would motivate them to engage with results and provide a comparison point for their answer. we hoped to learn about their satisfaction with relevance, quality, and quantity of results. (user behavior, user satisfaction) . using quick search find a video related to early childhood cognitive development. when you’ve found a suitable video recording, click answer and copy and paste the title. this task aimed to determine whether participants could complete the task, as well as show us which features they used in their attempts. (usability, user behavior) . search on speech pathology and find a way to limit your search results to audiology. then, limit your search results to peer reviewed sources. how satisfied are you with the results? since there are several ways to limit results in eds, we designed this task to show us which limiters participants tried to use, and which limiters resulted in success. we also hoped to learn about whether they thought the limiters provided satisfactory results. (usability, user behavior, user satisfaction) usability test results for a discovery tool in an academic library | fagan et al . you need more recent sources. please limit these search results to the last years, then select the most recent source available. click finished when you are done. since there are several ways to limit by date in eds, we designed this task to show us which limiters participants tried to use, and which limiters resulted in success. (usability, user behavior) . find a way to ask a jmu librarian for help using this search tool. after you’ve found the correct web page, click finished. we wanted to determine whether the user could complete this task, and which pathway they chose to do it. (usability, user behavior) . locate the journal yachting and boating world. what are the coverage dates? is this journal available in online full text? we wanted to determine whether the user could locate a journal by title. (usability) . you need to look up the sculpture genius of mirth. you have been told that the library database, camio, would be the best place to search for this. locate this database and find the sculpture. we wanted to know whether users who knew they needed to use a specific database could find that database from within the discovery tool. (usability, user behavior). . use quick search to find books and recent peer reviewed articles (from the last years) on rheumatoid arthritis. when you have found suitable source click answer and copy and paste the titles. click back to webpage if you need to return to your search results. these two tasks were intended to show us how users completed a common, broad task with and without a discovery tool, whether they would be more successful with or without the tool, and what barriers existed with and without the tool (usability, user behavior) . without using quick search, find books and recent peer reviewed articles (from the last years) on rheumatoid arthritis. when you have found suitable sources click answer and copy and paste the titles. click back to webpage if you need to return to your search results. editorial board thoughts: tools of the trade sharon farnel information technology and libraries | march as i was trying to settle on a possible topic for this, my second “editorial board thoughts” piece, i was struggling to find something that i’d like to talk about and that ital readers would (i hope) find interesting. i had my “eureka!” moment one day as i was coming out of a meeting, thinking about a conversation that had taken place around tools. now, by tools, i’m referring not to hardware, but to those programs and applications that we can and do use to make our work easier. the meeting was of our institutional repository team, and the tools discussion specifically focused on data cleanup and normalization, citation integration, and the like. i had just recently returned from a short conference where i had heard mentioned or seen demonstrated a few neat applications that i thought had potential. a colleague also had just returned from a different conference, excited by some of things that he’d learned about. and all of the team members had, in recent days, seen various e-mail messages about new tools and applications that might be useful in our environment. we mentioned and discussed briefly some of the tools that we planned to test. one of the tools had already been test driven by a couple of us, and looked promising; another seemed like it might solve several problems, and so was bumped up the testing priority list. during the course of the conversation, it became clear that each of us had a laundry list of tools that we wanted to explore at greater depth. and it also became clear that, as is so often the case, the challenge was finding the time to do so. as we were talking, my head was full of images of an assembly line, widgets sliding by so quickly that you could hardly keep up. i started thinking how you could stand there forever, overwhelmed by the variety and number of things flying by at what seemed like warp speed. alternatively, if you ever wanted to get anywhere, do anything, or be a part of it all, you just had to roll up your sleeves and grab something. the meeting drew to a close, and we all left with a sense that we needed to find a way of tackling the tools-testing process, of sharing what we learn and what we know, all in the hope of finding a set of tools that we, as a team, could become skilled with. i personally felt a little disappointed at not having managed to get around to all of the tools i’d earmarked for further investigation. but i also felt invigorated at the thought of being able to share the load of testing and researching. if we could coordinate ourselves, we might be able to test drive even more tools, increasing the sharon farnel (sharon.farnel@ualberta.ca) is metadata and cataloguing librarian, university of alberta, edmonton, alberta, canada. mailto:sharon.farnel@ualberta.ca editorial board thoughts | farnel likelihood we’d stumble on the few that would be just right! we’d taken furtive steps towards this in the past, but nothing coordinated enough to make it really stick and be effective. i started wondering how other individuals and institutions manage not only to keep up with all of the new and potentially relevant tools that appear at an ever-increasing pace, but more so how they manage to determine which they will become expert at and use going forward. (although i was excited at what we were thinking of doing, i was quite sure that others were likely far ahead of us in this regard!) it made me realize that at some point i—and we—need to stop being bystanders to the assembly line, watching the endless parade of tools pass us by. we need to simply grab on to a tool and take it for a spin. if it works for what we need, we stick with it. if it doesn’t, we put it back on the line, and grab a different one. but at some point we have to take a chance and give something a shot. we’ve decided on a few methods we’ll try for taking full advantage of the tool-rich environment in which libraries exist today. our metadata team has set up a “test bench,” a workstation that we can all use and share for trying new tools. a colleague is going to organize monthly brown-bag talks at which team members can demonstrate tools that they’ve been working with and that they think have potential uses in our work. and we’re also thinking of starting an informal, and public, blog, where we can post, among other things, about new tools we’ve tried or are trying, what we’re finding works and how, and what doesn’t and why. we hope these and other initiatives will help us all stay abreast or even slightly ahead of new developments, be flexible in incorporating new tools into our workflows when it makes the most sense, and in building skills and expertise that benefit us and that can be shared with others. so, i ask you, our ital readers, how do you manage the assembly line of tools? how do you gather information on them, and when do you decide to take one off and give it a whirl? how do you decide when something is worth keeping, or when something isn’t quite the right fit and gets placed back on the line? why not let us know by posting on the italica blog? or, even better, why not write about your experience and submit it to ital? we’re always on the lookout for interesting and instructional stories on the tools of our trade! http://ital-ica.blogspot.com/ copyright: regulation out of line with our digital reality? abigail j. mcdermott information technology and libraries | march abstract this paper provides a brief overview of the current state of copyright law in the united states, focusing on the negative impacts of these policies on libraries and patrons. the article discusses four challenges current copyright law presents to libraries and the public in general, highlighting three concrete ways intellectual property law interferes with digital library services and systems. finally, the author suggests that a greater emphasis on copyright literacy and a commitment among the library community to advocate for fairer policies is vital to correcting the imbalance between the interests of the public and those of copyright holders. introduction in july , the library community applauded when librarian of congress james h. billington announced new exemptions to the digital millennium copyright act (dmca). those with visual disabilities and the librarians who serve them can now circumvent digital rights management (drm) software on e-books to activate a read-aloud function. in addition, higher education faculty in departments other than film and media studies can now break through drm software to include high-resolution film clips in class materials and lectures. however, their students cannot, since only those who are pursuing a degree in film can legally do the same. that means that english students who want to legally include high-resolution clips from the critically acclaimed film sense and sensibility in their final projects on jane austin’s novel will have to wait another three years, when the librarian of congress will again review the dmca. the fact that these new exemptions to the dmca were a cause for celebration is one indicator of the imbalanced state of the copyright regulations that control creative intellectual property in this country. as the consumer-advocacy group public knowledge asserted, “we continue to be disappointed that the copyright office under the digital millennium copyright act can grant extremely limited exemptions and only every three years. this state of affairs is an indication that the law needs to be changed.” this paper provides a brief overview of the current state of u.s. copyright law, especially developments during the past fifteen years, with a focus on the negative impact these policies have had and will continue to have on libraries, librarians, and the patrons they serve. this paper does not provide a comprehensive and impartial primer on copyright law, a complex abigail j. mcdermott (ajmcderm@umd.edu) is graduate research associate, the information policy and access center (ipac), and masters candidate in library science, university of maryland, college park. mailto:ajmcderm@umd.edu copyright: regulation out of line with our digital reality | mcdermott and convoluted topic, instead identifying concerns about the effects an out-of-balance intellectual property system is having on the library profession, library services, and creative expression in our digital age. as with any area of public policy, the battles over intellectual property issues create an every fluctuating copyright environment, and therefore, this article is written to be current with policy developments as of october . finally, this paper recommends that librarians seek to better educate themselves about copyright law, and some innovative responses to an overly restrictive system, so that we can effectively advocate on our own behalf, and better serve our patrons. the state of u.s. copyright law copyright law is a response to what is known as the “progress clause” of the constitution, which charges congress with the responsibility “to promote the progress of science and the useful arts . . . to this end, copyright assures authors the right to their original expression, but encourages others to build freely upon the ideas and information conveyed by a work.” fair use, a statutory exception to u.s. copyright law, is a complex subject, but a brief examination of the principle gets to the heart of copyright law itself. when determining fair use, courts consider . the purpose and character of the use; . the nature of the copyrighted work; . the amount and substantiality of the portion used in relation to the copyrighted work as a whole; and . the effect of the use upon the potential market for the copyrighted work. while fair use is an “affirmative defense” to copyright infringement, invoking fair use is not the same as admitting to copyright infringement. teaching, scholarship, and research, as well as instances in which the use is not-for-profit and noncommercial, are all legitimate examples of fair use, even if fair use is determined on a case-by-case basis. despite the byzantine nature of copyright law, there are four key issues that present the greatest challenges and obstacles to librarians and people in general: the effect of the dmca on the principle of fair use; the dramatic extension of copyright terms codified by the sonny bono copyright term extension act; the disappearance of the registration requirement for copyright holders; and the problem of orphan works. the digital millennium copyright act (dmca) the dmca has been controversial since its passage in . title i of the dmca implements two world intellectual property organization (wipo) treaties that obligate member states to enforce laws that make tampering with drm software illegal. the dmca added chapter to the u.s. copyright act ( u.s.c. §§ – ), and it criminalized the trafficking of “technologies designed to circumvent access control devices protecting copyrighted material from unauthorized information technology and libraries | march copying or use.” while film studios, e-book publishers, and record producers have the right to protect their intellectual property from illegal pirating, the dmca struck a serious blow to the principle of fair use, placing librarians and others who could likely claim fair use when copying a dvd or pdf file in a catch- scenario. while the act of copying the file may be legal according to fair use, breaking through any drm technology that prevents that copying is now illegal. the sonny bono copyright term extension act while the copyright act of only provided authors and publishers with twenty-eight years of copyright protection, the sonny bono copyright term extension act of increased the copyright terms of all copyrighted works that were eligible for renewal in to ninety-five years after the year of the creator’s death. in addition, all works copyrighted on or after january , , now receive copyright protection for the life of the creator plus seventy years (or ninety-five years from the date of publication for works produced by multiple creators). jack valenti, former president of the motion picture association of american, was not successful in pushing copyright law past the bounds of the constitution, which mandates that copyright be limited, although he did try to circumvent this constitutional requirement by suggesting that copyright terms last forever less one day. the era of automatic copyright registration perhaps the most problematic facet of modern u.s. copyright law appears at first glance to be the most innocuous. the copyright act of did away with the registration requirement established by the copyright act of . that means that any creative work “fixed in any tangible medium of expression” is automatically copyrighted at the moment of its creation. that includes family vacation photos stored on a computer hard drive; they are copyrighted and your permission is required to use them. the previous requirement of registration meant authors and creators had to actively register their works, so anything that was not registered entered the public domain, replenishing that important cultural realm. now that copyright attaches at the moment an idea is expressed through a cocktail napkin doodle or an outline, virtually nothing new enters the public domain until its copyright term expires—at least seventy years later. in fact, nothing new will enter the public domain through copyright expiration until . until then, the public domain is essentially frozen in the year . the problem of orphan works in addition, the incredibly long copyright terms that apply to all books, photographs, and sound recordings have created the problem of orphan works. orphan works are those works that are under copyright protection, but whose owners are difficult or impossible to locate, often due to death. these publications are problematic for researchers, librarians, and the public in general: orphan works are perceived to be inaccessible because of the risk of infringement liability that a user might incur if and when a copyright owner subsequently appears. consequently, many works that are, copyright: regulation out of line with our digital reality | mcdermott in fact, abandoned by owners are withheld from public view and circulation because of uncertainty about the owner and the risk of liability. if copyright expired with the death of the author, or if there were a clause that would allow these works to pass into the public domain if the copyright holder’s heirs did not actively renew copyright for another term, then these materials would be far less likely to fall into legal limbo. currently, many are protected despite the fact that acquiring permission to use them is all but impossible. a study of orphan works in the collections of united kingdom public sector institutions found that these works are likely to have little commercial value, but high “academic and cultural significance,” and when contacted, these difficult-to-trace rights holders often grant permission for reproduction without asking for compensation. put another way, orphan works are essentially “locking up culture and other public sector content and preventing organizations from serving the public interest.” the row that arose in september between the hathitrust institutions and the authors guild over the university of michigan’s orphan works digitization project, with j. r. salamanca’s long- out-of-print novel the lost country serving as the pivot point in the dispute, is an example of the orphan works problem. the fact that university of michigan associate university librarian john price wilkin was forced to assure the public that “no copyrighted books were made accessible to any students” illustrates the absurdity in arguing over whether it’s right to digitize books that are no longer even accessible in their printed form. libraries, digitization, and copyright law: the quiet crisis while one can debate if u.s. copyright law is still oriented toward the public good, the more relevant question in this context is the effect copyright law has on the library profession. drm technology can get in the way of serving library patrons with visual disabilities and every library needs to place a copyright disclaimer on the photocopiers, but how much more of a stumbling block is intellectual property law to librarians in general, and the advance of library systems and technology in particular? the answer is undeniably that current u.s. copyright legislation places obstacles in the way of librarians working in all types of libraries. while there are many ways that copyright law affects library services and collections in this digital area, three challenges are particularly pressing: the problem of ownership and licensing of digital content or collections; the librarian as de facto copyright expert; and copyright law as it relates to library digitization programs generally, and the google book settlement in particular. digital collections: licenses replace ownership in the past, people bought a book, and they owned that copy. there was little they could accidentally or unknowingly do to infringe on the copyright holder’s rights. likewise, when physical collections were their only concern, librarians could rely on sections and of the copyright law to protect them from liability when they copied a book or other work and when they loaned materials in their collections to patrons. today, we live partly in the physical world and https://exch.mail.umd.edu/owa/webreadyviewbody.aspx?t=att&id=rgaaaadxslsgbeewtj q yhnkit bwboujgpo tvsou x% bkwiyfqalrqjtslaaboujgpo tvsou x% bkwiyfqapiuledyaaaj&attid =eacjse zzphuq qbfqvhbhu &attcnt= &pn= #footnote #footnote information technology and libraries | march partly in the digital world, reaching out and connecting to each other across fiber optic lines in the same way we once did around the water cooler. likewise, the digital means of production are widely distributed. in a multimedia world, where sharing an informative or entertaining video clip is as easy as embedding a link onto someone’s facebook wall, the temptation to infringe on rights by distributing, reproducing, or displaying a creative work is all too common, and all too easy. many librarians believe that disclaimers on public-access computer terminals will protect them from lawsuit, but they do not often consider placing such disclaimers on their cd or dvd collections. yet a copyright holder would not have to prove the library is aware of piracy to accuse the library of vicarious infringement of copyright. the copyright holder may even be able to argue that the library sees some financial gain from this piracy if the existence of the material that is being pirated serves as the primary reason a patron visits the library. even the physical cd collection in the public library can place the institution in danger of copyright infringement; yet the copyright challenges raised by cutting-edge digital resources, like e-books, are undoubtedly more complicated. e-books are replacing traditional books in many contexts. like most digital works today, e-books are licensed, not purchased outright. the problem licensing presents to libraries is that licensed works are not sold, they are granted through contracts, and contracts can change suddenly and negate fair-use provisions of u.s. copyright law. while libraries are now adept at negotiating contracts with subscription database providers, e-books are in many ways even more difficult to manage, with many vendors requiring that patrons delete or destroy the licensed content on their personal e-readers at the end of the lending period. the entire library community was rocked by harpercollins’s february decision to limit licenses on e-books offered through library e- book vendors like overdrive to twenty-six circulations, with many librarians questioning the publisher’s assertion that this seemingly arbitrary limitation is related to the average lifespan of a single print copy. license holders have an easy time arguing that any use of their content without paying fees is a violation of their copyright. that is not the case when a fair use argument is justified, and while many in the library community may acquiesce to these arguments, “in recent cases, courts have found the use of a work to be fair despite the existence of a licensing market.” when license agreements are paired with drm technology, libraries may find themselves managing thousands of micropayments to allow their users to view, copy, move, print, or embed, for example, the pdf of a scholarly journal article. in the current climate of reduced staff and shrinking budgets, managing these complex licensing agreements has the potential to cripple many libraries. the librarian as accidental copyright czar during a special libraries association (sla) q&a session on copyright law in the digital age, the questions submitted to the panel came from librarians working in hospitals, public libraries, academic libraries, and even law libraries. librarians are being thrust into the position of de facto copyright expert. one of the speakers mentioned that she must constantly remind the lawyers at copyright: regulation out of line with our digital reality | mcdermott the firm she works for that they should not copy and paste the full text of news or law review journal articles into their e-mails, and instead, they should send a link. the basis of her argument is the third factor of fair use mentioned earlier: the amount or substantiality of the portion of the copyrighted work being used. since fair use is not a “bright line” principle, the more factors you have on your side the better when you are using a copyrighted work without the owners express permission. librarians working in any institution must seek express permission from copyright holders for any video they wish to post, or embed, on library-managed websites. e-reserves and streaming video, mainstays of many educators and librarians seeking to capture the attention of this digital generation, have become bright red targets for litigious copyright holders who want to shrink the territory claimed under the fair-use banner even further. many in the library community are aware of the georgia state university e-reserves lawsuit, cambridge university press et al. v. patton, in which a group of academic publishers have accused the school of turning its e-reserves system into a vehicle for intentional piracy. university librarians are implicated for not providing sufficient oversight. it has come to light that the association of american publishers (aap) approached other schools, including cornell, hofstra, syracuse, and marquette, before filing a suit against georgia state. generally, the letters come from aaps outside counsel and are accompanied by “the draft of a federal court legal complaint that alleges copyright infringement.” the aap believes that e-reserves are by nature an infringement of copyright law, so they demand these universities work with their association to draft guidelines for electronic content that support aaps “cost-per-click theory of contemporary copyright: no pay equals no click.” it seems that georgia state was not willing to quietly concede to aap’s view on the matter, and they are now facing the association in court. a decision in this case was pending at the time this article went to press. the case brought by the association for information and media equipment (aime) against ucla is similar, except it focuses on the posting of videos so they can be streamed by students on password-protected university websites that do not allow the copying or retention of the videos. ucla argued that the video streaming services for students are protected by the technology education and copyright harmonization (teach) act of , which is the same act that allows all libraries to offer patrons online access to electronic subscription databases off-site through a user-authentication system. in addition, ucla argues that it is simply allowing its students to “time shift” these videos, a practice deemed not to infringe on copyright law by the supreme court in its landmark sony corp. v. universal city studios, inc. decision of . the american library association (ala), association of research libraries (arl), and the association of college and research libraries (acrl) jointly published an opinion supporting ucla in this case. many in the wider library community sympathized with ucla’s library administrators, who cite budget cuts that reduced hours at the school’s media laboratory as one reason they must now offer students a video-streaming option. in the end, the case was dismissed, mostly due to the lack of standing aime had to bring the suite against ucla, a state agency, in federal court. while the judge did not https://exch.mail.umd.edu/owa/webreadyviewbody.aspx?t=att&id=rgaaaadxslsgbeewtj q yhnkit bwboujgpo tvsou x% bkwiyfqalrqjtslaaboujgpo tvsou x% bkwiyfqapiuledyaaaj&attid =eacjse zzphuq qbfqvhbhu &attcnt= &pn= #footnote #footnote information technology and libraries | march expressly rule on the fair-use argument ucla made, the ruling did confirm that streaming is not a form of video distribution and that the public-performance argument ucla made regarding the videos was not invalidated by the fact that they made copies of the videos in question. digitization programs and the google book settlement librarians looking to digitize print collections, either for preservation or to facilitate online access, are also grappling with the copyright monopoly. librarians who do not have the time or resources to seek permission from publishers and authors before scanning a book in their collection cannot touch anything published after . librarylaw.com provides a helpful chart directed at librarians considering digitization projects, but the overwhelming fine print below the chart speaks to the labyrinthine nature of copyright. the google book settlement continues to loom large over both the library profession and the publishing industry. at the heart of debate is google’s library project, which is part of google book search, originally named google print. the library project allows users to search for books using google’s algorithms to provide at its most basic a “snippet view” of the text from a relevant publication. authors and publishers could also grant their permission to allow a view of select sample pages, and of course if the book is in the public domain, then google can make the entire work visible online. in all cases, the user will see a “buy this book” link so that he or she could purchase the publication from online vendors on unrelated sites. google hoped to sidestep the copyright permission quandary for a digitization project of this scale, announcing that it would proceed with the digitization of cooperative library collections and that it would be the responsibility of publishers and authors to actively opt out or vocalize their objection to seeing their works digitized and posted online. google attempted to turn the copyright permissions process on its head, which was the basis of the class action lawsuit authors guild v. google inc. before the settlement was reached, google pointed to kelly v. arriba soft corp as proof that the indexing functions of an internet search engine constitute fair use. in that case, the ninth circuit court of appeals found that a website’s posting of thumbnail images, or “imprecise copies of low resolution, scaled down images,” constitutes fair use, and google argued its “snippet view” function is equivalent to a thumbnail image. however, judge denny chin rejected the google book settlement in march , citing the fact that google would in essence be “exploiting books without the permission of copyright owners” and could also establish a monopoly over the digitized books market. the decision did in the end hinge on the fact that google wanted to follow an opt-out program for copyright holders rather than an affirmative opt-in system. the google book settlement was dismissed without prejudice, leaving the door open to further negotiations between the parties concerned. going forward, the library community should be concerned with how google will handle orphan works and how its index of digitized works will be made available to libraries and the public. the settlement granted google the nonexclusive right to digitize all books published before january , , and in exchange, google would have https://exch.mail.umd.edu/owa/webreadyviewbody.aspx?t=att&id=rgaaaadxslsgbeewtj q yhnkit bwboujgpo tvsou x% bkwiyfqalrqjtslaaboujgpo tvsou x% bkwiyfqapiuledyaaaj&attid =eacjse zzphuq qbfqvhbhu &attcnt= &pn= #footnote #footnote copyright: regulation out of line with our digital reality | mcdermott “paid % of the net revenue earned from uses of google book search in the united states to rights holders.” in addition, google would have established the book rights registry to negotiate with google and others seeking to “digitize, index or display” those works on behalf of the rights holders. approval of the settlement would have allowed google to move forward with plans to expand google book search and “to sell subscriptions to institutions and electronic versions of books to individuals.” the concern that judge denny chin expressed over a potential google book monopoly was widespread among the library community. while the settlement would not have given google exclusive rights to digitize and display these copyrighted works, google planned to ensure via the settlement that it would have received the same terms the book rights registry negotiated with any third-party digital library, while also inoculating itself against the risk of any copyright infringement lawsuits that could be filed against a competitor. that would have left libraries vulnerable to any subscription price increases for the google books service. libraries should carefully watch the negotiations around any future google books settlement, paying attention to a few key issues. there was considerable concern that under the terms of the settlement, even libraries participating in the google books library project would need to subscribe to the service to have access to digitized copies of the books in their own collections. many librarians also vocalized their disappointment in google’s abandonment of its fair-use argument when it agreed to the settlement, which, if it succeeded, would have been a boon to nonprofit, library-driven digitization programs. finally, many librarians were concerned that google’s book rights registry was likely to become the default rights holder for the orphan works in the google books library, and that claims that google books is an altruistic effort to establish a world library conceals the less admirable aim of the project—to monetize out-of-print and orphan works. librarians as free culture advocates: implications and recommendations our digital nation has turned copyright law into a minefield for both librarians and the public at large. intellectual property scholar lawrence lessig failed in his attempt to argue before the supreme court that the sonny bono copyright term extension act was an attempt to regulate free speech and therefore violated the first amendment. but many believe that our restrictive copyright laws at least violate the intent of the progress clause of the constitution, if not the first amendment: “unconstrained access to past works helps determine the richness of future works. inversely, when past works are inaccessible except to a privileged minority, future works are impoverished.” while technological advances have placed the digital means of production into the hands of the masses, intellectual property law is leading us down a path to self-censorship. as the profession “at the heart of both the knowledge economy and a healthy democracy,” it is in our best interest as librarians to recognize the important role we have to play in restoring the balance to copyright law. to engage in the debate over copyright law in the digital age, the library community needs to educate itself and advocate for our own self-interests, focusing on three key areas: https://exch.mail.umd.edu/owa/webreadyviewbody.aspx?t=att&id=rgaaaadxslsgbeewtj q yhnkit bwboujgpo tvsou x% bkwiyfqalrqjtslaaboujgpo tvsou x% bkwiyfqapiuledyaaaj&attid =eacjse zzphuq qbfqvhbhu &attcnt= &pn= #footnote #footnote information technology and libraries | march . copyright law in the classroom and at the conference. we must educate new and seasoned librarians on the nature of copyright law, and the impact it has on library practice and systems. library schools must step up to the plate and include a thorough overview of copyright law in their library science curriculum. while including copyright law in a larger legal-issues class is acceptable, the complexity of current u.s. copyright law demonstrates that this is not a subject that can be glossed over in a single lecture. furthermore, there needs to be a stronger emphasis on continuing education and training on copyright law within the library profession. the sla offers a copyright certificate program, but the reach of such programs is not wide enough. copyright law, and the impacts current policy has on the library profession, must be prominently featured at library conferences. the university of maryland university college’s center for intellectual property offers an online community forum for discussing copyright issues and policies, but it is unclear how many librarians are members. . librarians as standard-bearers for the free culture movement. while the library copyright alliance, to which the ala, arl, and acrl all belong, files amicus briefs in support of balanced copyright law and submits comments to wipo, the wider library community must also advocate for copyright reform, since this is an issue that affects all librarians, everywhere. as a profession, we need to throw our collective weight behind legislative measures that address the copyright monopoly. there have been a number of unfortunate failures in recent years. s. , or the consumers, schools, and libraries digital management awareness act of , attempted to address a number of drm issues, including a requirement that access controlled digital media and electronics include disclosures on the nature of the drm technology in use. h.r. , the digital media consumers rights act of , would have amended the dmca to allow those researching the technology to circumvent drm software while also eliminating the catch- that makes circumventing drm software for fair-use purposes illegal. the balance act of (h.r. ) included provisions to expand fair use to the act of transmitting, accepting, and saving a copyrighted digital work for personal use. all of this legislation died in committee, as did h.r. (orphan works act of ) and s. (shawn bentley orphan works act of ). both bills would have addressed the orphan works dilemma, clearly spelling out the steps one must take to use an orphan work with no express permission from the copyright holder, without fear of a future lawsuit. could a show of support from the library community have saved these bills? it is impossible to know, but it is in our best interest to follow these legislative battles in the future and make sure our voice is heard. . libraries and the creative commons phenomenon. in addition, librarians need to take part in the creative commons (cc) movement by actively directing patrons towards this world of digital works that have clear, simple use and attribution requirements. creative commons was founded in with the support of the center for the study of the public domain at duke university school of law. the movement is essentially about free culture, and the idea that many people want to share their creative works and allow others to use or build off of their efforts easily and without seeking their permission. it is not intended to supplant copyright law, and lawrence copyright: regulation out of line with our digital reality | mcdermott lessing, one of the founders of creative commons, has said many times that he believes intellectual property law is necessary and that piracy is inexcusable. instead, a cc license states in clear terms exactly what rights the creator reserves, and conversely, what rights are granted to everyone else. as lawrence lessig explains, you go to the creative commons website (http://creativecomms.org); you pick the opportunity to select a license: do you want to permit commercial uses or not? do you want to allow modifications or not? if you allow modifications, do you want to require a kind of copyleft idea that other people release the modifications under a similarly free license? that is the core, and that produces a license. there are currently six cc licenses, and they include some combination of the four license conditions defined by creative commons: attribution (by), share alike (sa), noncommercial (nc), and no derivatives (nd). each of the four conditions is designated by a clever symbol, and the six licenses display these symbols after the creative commons trademark itself, two small c’s inside a circle. there are “hundreds of millions of cc licensed works” that can be searched through google and yahoo, and some notable organizations that rely on cc licenses include flickr, the public library of science, wikipedia, and now whitehouse.gov. all librarians not already familiar with this approach need to educate themselves on cc licenses and how to find cc licensed works. while librarians must still inform their patrons about the realities of copyright law, it is just as important to direct patrons, students, and colleagues to cc licensed materials, so that they can create the mash-ups, videos, and podcasts that are the creative products of our web . world. the creative commons system is not perfect, and “creative commons gives the unskilled an opportunity to fail at many junctures.” yet that only speaks to the necessity of educating the library community about the “some rights reserved” movement, so that librarians, who are already called upon to understand traditional copyright law, are also educating our society about how individuals can protect their intellectual property while preserving and strengthening the public domain. conclusion the library community can no longer afford to consider intellectual property law as a foreign topic appropriate for law schools but not library schools. those who are behind the slow extermination of the public domain rely on the complexity of copyright law, and the misunderstanding of the principle of fair use, to make their arguments easier and to brow beat libraries and the public into handing over the rights the constitution bestows on everyone. librarians need to engage in the debate over copyright law to retain control over their collections, and to better serve their patrons. in the past, the library community has not hesitated to stand up for the freedom of speech and self-expression, whether it means taking a stand against banning books from school libraries or fighting to repeal clauses of the usa patriot act. today’s library patrons are not just information consumers—they are also information producers. therefore it is just as critical for librarians to advocate for their creative rights as it is for them to defend their freedom to read. https://exch.mail.umd.edu/owa/webreadyviewbody.aspx?t=att&id=rgaaaadxslsgbeewtj q yhnkit bwboujgpo tvsou x% bkwiyfqalrqjtslaaboujgpo tvsou x% bkwiyfqapiuledyaaaj&attid =eacjse zzphuq qbfqvhbhu &attcnt= &pn= #footnote #footnote https://exch.mail.umd.edu/owa/redir.aspx?c= cfbeb fb d b edf dcfd&url=http% a% f% fcreativecomms.org https://exch.mail.umd.edu/owa/webreadyviewbody.aspx?t=att&id=rgaaaadxslsgbeewtj q yhnkit bwboujgpo tvsou x% bkwiyfqalrqjtslaaboujgpo tvsou x% bkwiyfqapiuledyaaaj&attid =eacjse zzphuq qbfqvhbhu &attcnt= &pn= #footnote #footnote information technology and libraries | march the internet has become such a strong incubator of creative expression and innovation that the innovators are looking for a way to shirk the very laws that were designed to protect their interests. in the end, the desire to create and innovate seems to be more innate than those writing our intellectual property laws expected. perhaps financial gain is less of a motivator than the pleasure of sharing a piece of ourselves and our worldview with the rest of society. whether that’s the case or not, what is clear is that if we do not roll back legislation like the sonny bono copyright term extension act and the dmca so as to save the public domain, the pressure to create outside the bounds of the law is going to turn more inventors and artists into anarchists, threatening the interests of reasonable copyright holders. as librarians, we must curate and defend the creative property of the established, while fostering the innovative spirit of the next generation. as information, literature, and other creative works move out of the physical world, and off the shelves, into the digital realm, librarians need to do their part to ensure legislation is aligned with this new reality. if we do not, our profession may suffer first, but it will not be the last casualty of the copyright wars. references . beverly goldberg, “lg unlocks doors for creators, consumers with dmca exceptions,” american libraries , no. (summer ): . . ibid. . goldberg, “lg unlocks doors.” . christopher alan jennings, fair use on the internet, prepared by the congressional research service (washington, dc: library of congress, ), . . ibid., . . ibid. . brandon butler, “urban copyright legends,” research library issues (june ): . . robin jeweler, “digital rights” and fair use in copyright law, prepared by the congressional research service (washington, dc: library of congress, ), . . rachel bridgewater, “tipping the scales: how free culture helps restore balance in the age of copyright maximalism,” oregon library association quarterly , no. (fall ): . . charles w. bailey jr., “strong copyright + drm + weak net neutrality = digital dystopia?” information technology & libraries , no. (summer ): ; u.s. copyright office, “copyright law of the united states,” under “chapter : duration of copyright,” http://www.copyright.gov/title (accessed december , ). . dan hunter, “culture war,” texas law review , no. ( ): . . bailey, “strong copyright,” . . u.s. copyright office, “copyright law of the united states,” under “chapter : subject matter and scope of copyright,” http://www.copyright.gov/title (accessed december , ). . bailey, “strong copyright,” . . mary minnow, “library digitization table,” http://www.librarylaw.com/digitizationtable.htm (accessed december , ). https://exch.mail.umd.edu/owa/redir.aspx?c= cfbeb fb d b edf dcfd&url=http% a% f% fwww.copyright.gov% ftitle https://exch.mail.umd.edu/owa/redir.aspx?c= cfbeb fb d b edf dcfd&url=http% a% f% fwww.copyright.gov% ftitle https://exch.mail.umd.edu/owa/redir.aspx?c= cfbeb fb d b edf dcfd&url=http% a% f% fwww.librarylaw.com% fdigitizationtable.htm copyright: regulation out of line with our digital reality | mcdermott . brian t. yeh, “orphan works” in copyright law, prepared by the congressional research service (washington, dc: library of congress, ), summary. . ibid. . jisc, in from the cold: an assessment of the scope of “orphan works” and its impact on the delivery of services to the public (cambridge, uk: jisc, ), . . ibid. . andrew albanese, “hathitrust suspends its orphan works release,” publishers weekly, sept, , , http://www.publishersweekly.com/pw/by- topic/digital/copyright/article/ -hathitrust-suspends-its-orphan-works-release-.html (accessed october , ). . u.s. copyright office, “copyright law of the united states,” under “chapter .” . u.s. copyright office, copyright basics (washington, dc: u.s. copyright office, ), www.copyright.gov/circs/circl/html (accessed december , ). . mary minnow, california library association, “library copyright liability and pirating patrons,” http://www.cla-net.org/resources/articles/minow_pirating.php (accessed december , ). . bailey, “strong copyright,” . . overdrive, “copyright,” http://www.overdrive.com/copyright.asp (accessed december , ). . josh hadro, “harpercollins puts loan cap on ebook circulations,” library journal (february ), http://www.libraryjournal.com/lj/home/ - /harpercollins_puts_ _loan_cap.html.csp (accessed october , ). . butler, “urban copyright legends,” . . bailey, “strong copyright,” . . library of congress, fair use on the internet, . . ibid., summary. . matthew k. dames, “education use in the digital age,” information today , no. (april ): . . ibid. . dames, “education use in the digital age,” . . matthew k. dames, “making a case for copyright officers,” information today , no. (july ): . . william c. dougherty, “the copyright quagmire,” journal of academic librarianship , no. (july ): . . ibid. . library of congress, “digital rights” and fair use in copyright law, . . dougherty, “the copyright quagmire,” . . kevin smith, “streaming video case dismissed,” scholarly communications @ duke, october , , http://blogs.library.duke.edu/scholcomm/ / / /streaming-video-case- dismissed/ (accessed october , ). http://www.publishersweekly.com/pw/by-topic/digital/copyright/article/ -hathitrust-suspends-its-orphan-works-release-.html http://www.publishersweekly.com/pw/by-topic/digital/copyright/article/ -hathitrust-suspends-its-orphan-works-release-.html https://exch.mail.umd.edu/owa/redir.aspx?c= cfbeb fb d b edf dcfd&url=http% a% f% fwww.copyright.gov% fcircs% fcircl% fhtml https://exch.mail.umd.edu/owa/redir.aspx?c= cfbeb fb d b edf dcfd&url=http% a% f% fwww.cla-net.org% fresources% farticles% fminow_pirating.php https://exch.mail.umd.edu/owa/redir.aspx?c= cfbeb fb d b edf dcfd&url=http% a% f% fwww.overdrive.com% fcopyright.asp http://www.libraryjournal.com/lj/home/ - /harpercollins_puts_ _loan_cap.html.csp http://www.libraryjournal.com/lj/home/ - /harpercollins_puts_ _loan_cap.html.csp http://blogs.library.duke.edu/scholcomm/ / / /streaming-video-case-dismissed/ http://blogs.library.duke.edu/scholcomm/ / / /streaming-video-case-dismissed/ information technology and libraries | march . dougherty, “the copyright quagmire,” . . librarylaw.com, “library digitization table.” . kate m. manuel, the google library project: is digitization for purposes of online indexing fair use under copyright law, prepared by the congressional research service (washington, dc: library of congress, ), – . . jeweler, “digital rights” and fair use in copyright law, . . ibid. . ibid. . manuel, the google library project, . . amir efrati and jeffrey a. trachtenberg, “judge rejects google books settlement,” wall street journal, march , , http://online.wsj.com/article/sb .html (accessed october , ). . jennings, fair use on the internet, . . manuel, the google library project, . . ibid., – . . ibid. . ibid. . pamela samuelson, “google books is not a library,” huffington post, october , , http://www.huffingtonpost.com/pamela-samuelson/google-books-is-not-a-lib_b_ .html (accessed december , ). . ivy anderson, “hurtling toward the finish line: should the google book settlement be approved?” against the grain , no. (june ): . . samuelson, “google books is not a library.” . jeweler, “‘digital rights” and fair use in copyright law, . . bailey, “strong copyright,” . . cushla kapitzke, “rethinking copyrights for the library through creative commons licensing,” library trends , no. (summer ): . . ibid. . university of maryland university college, “member community,” center for intellectual property, http://cipcommunity.org/s/ /start.aspx (accessed february , ). . robin jeweler, copyright law: digital rights management legislation, prepared by the congressional research service (washington, dc: library of congress, ), summary. . creative commons, “history,” http://creativecommons.org/about/history/ (accessed december , ). . lawrence lessig, “the vision for the creative commons? what are we and where are we headed? free culture,” in open content licensing: cultivating the creative commons, ed. brian fitzgerald (sydney: sydney university press, ), . . steven j. melamut, “free creativity: understanding the creative commons licenses,” american association of law libraries , no. (april ): . http://online.wsj.com/article/sb .html https://exch.mail.umd.edu/owa/redir.aspx?c= cfbeb fb d b edf dcfd&url=http% a% f% fwww.huffingtonpost.com% fpamela-samuelson% fgoogle-books-is-not-a-lib_b_ .html https://exch.mail.umd.edu/owa/redir.aspx?c= cfbeb fb d b edf dcfd&url=http% a% f% fcipcommunity.org% fs% f % fstart.aspx https://exch.mail.umd.edu/owa/redir.aspx?c= cfbeb fb d b edf dcfd&url=http% a% f% fcreativecommons.org% fabout% fhistory% f copyright: regulation out of line with our digital reality | mcdermott . lessig, “the vision for the creative commons?” . . creative commons, “about,” http://creativecommons.org/about/ (accessed december , ). . ibid. . ibid. . bridgewater, “tipping the scales,” . . ibid. . woody evans, “commons and creativity,” searcher , no. (october ): . https://exch.mail.umd.edu/owa/redir.aspx?c= cfbeb fb d b edf dcfd&url=http% a% f% fcreativecommons.org% fabout% f batch ingesting into eprints digital repository sof tware tomasz neugebauer and bin han information technology and libraries | march abstract this paper describes the batch importing strategy and workflow used for the import of theses metadata and pdf documents into the eprints digital repository software. a two-step strategy of importing metadata in marc format followed by attachment of pdf documents is described in detail, including perl source code for scripts used. the processes described were used in the ingestion of , theses metadata and pdfs into an eprints institutional repository. introduction tutorials have been published about batch ingestion of proquest metadata and electronic theses and dissertations (etds), as well as endnote library, into the digital commons platform. the procedures for bulk importing of etds using dspace have also been reported. however, bulk importing into the eprints digital repository software has not been exhaustively addressed in the literature. a recent article by walsh provides a literature review of batch importing into institutional repositories. the only published report on batch importing into the eprints platform describes perl scripts for metadata-only records import from thomson reuters reference manager. bulk importing is often one of the first tasks after launching a repository, so it is unsurprising that requests for reports and documentation on eprints-specific workflow have been a recurring question on the eprints tech list. a recently published review of eprints identifies “the absence of a bulk uploading feature” as its most significant weakness. although eprints’ graphical user interface for bulk importing is limited to the use of the installed import plugins, the software does have a versatile infrastructure for this purpose. leveraging eprints’ import functionality requires some perl scripting, structuring the data for import, and using the command line interface. in , when concordia university launched spectrum, its research repository, the first task was a batch ingest of approximately , theses dated from to . the source of the metadata for this import consisted in marc records from an integrated library system powered by innovative interfaces and proquest pdf documents. this paper is a report on the strategy and workflow adopted for batch ingestion of this content into the eprints digital repository software. import strategy eprints has a documented import command line utility located in the /bin folder. documents can also be imported through eprints’ graphical interface. using the command line utility for tomasz neugebauer (tomasz.neugebauer@concordia.ca) is digital projects and systems development librarian and bin han (bin.han@concordia.ca) is digital repository developer, concordia university libraries, montreal, quebec, canada. mailto:tomasz.neugebauer@concordia.ca mailto:bin.han@concordia.ca batch ingesting into eprints digital repository software| neugebauer and han importing is recommended because it is easier to monitor the operation in real time by adding progress information output to the import plugin code. the task of batch importing can be split into the following subtasks: . import of metadata of each item . import of associated documents, such as full-text pdf files the strategy adopted was to first import the metadata for all of the new items into the inbox of an editor’s account. after this first step was completed, a script was used to loop through the newly imported eprints and attach the corresponding full-text documents. although documents can be imported from the local file system or via http, import of the files from the local file system was used. the batch import procedure varies depending on the format of the metadata and documents to be imported. metadata import requires a mapping of the source schema fields to the default or custom fields in eprints. the source metadata must also be converted into one of the formats supported by eprints’ import plugins, or a custom plugin must be created. import plugins are available for many popular formats, including bibtex, doi, endnote, and pubmedxml. in addition, community-contributed import plugins such as marc and arxiv are available at eprints files. since most repositories use custom metadata fields, some customization of the import plugins is usually necessary. marc plugin for eprints in eprints, the import and export plugins ensure interoperability of the repository with other systems. import plugins read metadata from one schema and load it into the eprints system through a mapping of the fields into the eprints schema. loading marc-encoded files into eprints requires the installation of the import/export plugin developed by romero and miguel. the installation of this plugin requires the following two cpan modules: marc::record and marc::file::usmarc. the marc plugin was then subclassed to create an import plugin named “concordia theses,” which is customized for thesis marc records. concordia theses marc plugin the marc plugin features a central configuration file (see appendix a) in which each marc field is paired with a corresponding mapping to an eprints field. most of the fields were configured through this configuration file (see table ). the source marc records from the innovative interfaces integrated library system (ils) encode the physical description of each item using the anglo american cataloguing rules, as in the following example: “ix, leaves : ill. ; cm.” since the default eprints field for number of pages is of the type integer and does not allow multipart physical descriptions from the marc field, a custom text field for these physical descriptions (pages_aacr) had to be added. the marc.pl configuration file cannot be used to map compound fields, such as author names—the fields need custom mapping implementation in perl. for instance, the marc and fields information technology and libraries | march are transferred into the eprints author compound field (in marc.pm). similarly, marc is mapped into a custom thesis advisor compound field. marc field eprints field a isbn z isbn a issn a title a edition a place_of_pub b publisher c date a pages_aacr a volume a series c volume x issn a abstract a publication table . mapping table from marc to eprints helge knüttel’s refinements to the marc plugin shared on the eprints tech list were employed in the implementation of a new subclass of marc import for the concordia theses marc records. in the implementation of the concordia theses plugin, concordiatheses.pm inherits from marc.pm. (see figure .) knüttel added two methods that make it easier to subclass the general marc plugin and add unique mappings: handle_marc_specialities and post_process_eprint. the post_process_eprint function was not used to attach the full-text documents to each eprint. instead, the strategy to import the full-text documents using a separate attach_documents script was used (see “theses document file attachment” below). import of all of the specialized fields, such as thesis type (mapped from marc t), program, department, and proquest id, was implemented in the function handle_marc_specialities of concordiatheses.pm. for instance, a in the marc record contains the department information, whereas an eprints system like spectrum stores department hierarchy as subject objects in a tree. therefore importing the department information based on the value of a required regular expression searches of this marc field to find the mapping into a corresponding subject id. this was implemented in the handle_marc_specialities function. batch ingesting into eprints digital repository software| neugebauer and han figure . concordia theses class diagram, created with the perl module uml::class::simple execution of the theses metadata import the depositing user’s name is displayed along with the metadata for each eprint. a batchimporter user with the corporate name “concordia university libraries” was created to carry out the import. as a result, the public display of the imported items shows the following as a part of the metadata: “deposited by: concordia university libraries.” the marc plugin requires the encoding of the source marc files to be utf- , whereas the records are exported from the ils with marc- encoding. therefore marcedit software developed by reese was used to convert the marc file to utf- . to activate the import, the main marc import plugin and its subclass, concordiatheses.pm, have to be placed in the plugin folder /perl_lib/eprints/plugin/import/marc/. the configuration file information technology and libraries | march (see appendix a) must also be placed with the rest of the configurable files in /archives/repositoryid/cfg/cfg.d. the plugin can then be activated from the command line using the import script in the /bin folder. a detailed description of this script and its usage is documented on the eprints wiki. the following eprints command from the /bin folder was used to launch the import: import repositoryid --verbose --user batchimporter eprint marc::concordiatheses theses-utf .mrc following the aforementioned steps, all the theses metadata was imported into the eprints software. the new items were imported with their statuses set to inbox. a status set to inbox means that the imported items are in the work area of batchimporter user and will need to be moved to live public access by switching their status to archive. theses document file attachment after the process of importing the metadata of each thesis is complete, the corresponding document files need to be attached. the proquest id was used to link the full-text pdf documents to the metadata records. all of the marc records contained the proquest id, while the pdf files, received from proquest, were delivered with the corresponding proquest id as the filename. the pdfs were uploaded to a folder on the repository web server using ftp. the attach_documents script (see appendix b for source code) was then used to attach the documents to each of the imported eprints in the batchimporter’s inbox and to move the imported eprints to the live archive. several variables need to be set at the beginning of the attach_documents operation (see table ). variable comment $root_dir = 'bin/import- data/proquest' this is the root folder where all the associated documents are uploaded by ftp. $depositor = 'batchimporter' only the items deposited by a defined depositor, in this case batchimporter, will be moved from inbox to live archive. $dataset_id = 'inbox' limit the dataset to those eprints with status set to inbox $repositoryid = 'library' the internal eprints identifier of the repository table . variables to be set in the attach_documents script batch ingesting into eprints digital repository software| neugebauer and han the following command is used to proceed with file attachment, while the output log is redirected and saved in the file attachment: /bin/attach_documents.pl > ./attachment >& the thesis metadata record was made live even if it did not contain a corresponding document file. a list of eprint ids of theses that did not contain a corresponding full-text pdf document are listed at the end of the log file, along with the count of the number of theses that were made live. after the import operation is complete, all the abstract pages need to be regenerated with the following command: /bin/generate_abstracts repositoryid conclusions this paper is a detailed report on batch importing into the eprints system. the authors believe that this paper and its accompanying source code is a useful contribution to the literature on batch importing into digital repository systems. in particular, it should be useful to institutions that are adopting the eprints digital repository software. batch importing of content is a basic and fundamental function of a repository system, which is why the topic has come up repeatedly on the eprints tech list and in a repository software review. the methods that we describe for carrying out batch importing in eprints make use of the command line and require perl scripting. more robust administrative graphical user interface support for batch import functions would be a useful feature to develop in the platform. acknowledgements the authors would like thank mia massicotte for exporting the metadata records from the integrated library system. we would also like to thank alexandros nitsiou, raquel horlick, adam field, and the reviewers at information technology and libraries for their useful comments and suggestions. references . shawn averkamp and joanna lee, “repurposing proquest metadata for batch ingesting etds into an institutional repository,” code{ }lib journal ( ), http://journal.code lib.org/articles/ (accessed june , ). . michael witt and mark p. newton, “preparing batch deposits for digital commons repositories,” , http://docs.lib.purdue.edu/lib_research/ / (accessed june , ). . randall floyd, “automated electronic thesis and dissertations ingest,” , https://wiki.dlib.indiana.edu/display/iusw/automated+electronic+thesis+and+dissertations+i ngest (accessed may , ). . eprints digital repository software, university of southampton, uk, http://www.eprints.org/ (accessed june , ). . maureen p. walsh, “batch loading collections into dspace: using perl scripts for automation and quality control,” information technology & libraries , no. ( ): – , http://journal.code lib.org/articles/ http://docs.lib.purdue.edu/lib_research/ / https://wiki.dlib.indiana.edu/display/iusw/automated+electronic+thesis+and+dissertations+ingest https://wiki.dlib.indiana.edu/display/iusw/automated+electronic+thesis+and+dissertations+ingest http://www.eprints.org/ information technology and libraries | march http://search.ebscohost.com/login.aspx?direct=true&db=a h&an= &site=ehost-live (accessed june , ). . lesley drysdale, “importing records from reference manager into gnu eprints,” , http://hdl.handle.net/ / (accessed june , ). . eprints tech list, university of southampton, uk, http://www.eprints.org/tech.php/ (accessed june , ). . mike beazly, “eprints institutional repository software: a review,” partnership: the canadian journal of library & information practice & research , no. ( ), http://journal.lib.uoguelph.ca/index.php/perj/article/viewarticle/ (accessed june , ). . concordia university libraries, “spectrum: concordia university research repository,” http://spectrum.library.concordia.ca (accessed june , ). . eprints wiki, “api:bin/import,” university of southampton, uk, http://wiki.eprints.org/w/api:bin/import (accessed june , ). . eprints files, university of southampton, uk, http://files.eprints.org/ (accessed june ). . parella romero and jose miguel, “marc import/export plugins for gnu eprints ,” eprints files, , http://files.eprints.org/ / (accessed may , ). . agent zhang and maxim zenin, “uml:class::simple,” cpan, http://search.cpan.org/~agent/uml-class-simple- . /lib/uml/class/simple.pm (accessed september , ). . terry reese, “marcedit: downloads,” oregon state university, http://people.oregonstate.edu/~reeset/marcedit/html/downloads.html (accessed june , ). http://search.ebscohost.com/login.aspx?direct=true&db=a h&an= &site=ehost-live http://hdl.handle.net/ / http://www.eprints.org/tech.php/ http://journal.lib.uoguelph.ca/index.php/perj/article/viewarticle/ http://spectrum.library.concordia.ca/ http://wiki.eprints.org/w/api:bin/import http://files.eprints.org/ http://files.eprints.org/ / http://search.cpan.org/~agent/uml-class-simple- . /lib/uml/class/simple.pm http://people.oregonstate.edu/~reeset/marcedit/html/downloads.html batch ingesting into eprints digital repository software| neugebauer and han appendix a. marc.pl configuration file # # plugin eprints::plugin::import::marc # # marc tofro eprints mappings # do _not_ add compound mappings here. $c->{marc}->{marc ep} = { # marc to eprints ' a' => 'isbn', ' z' => 'isbn', ' a' => 'issn', ' a' => 'title', ' b' => 'subtitle', ' a' => 'edition', ' a' => 'place_of_pub', ' b' => 'publisher', ' c' => 'date', ' a' => 'volume', ' a' => 'series', ' c' => 'volume', ' x' => 'issn', ' a' => 'abstract', ' a' => 'publication', }; $c->{marc}->{marc ep}->{constants} = { }; ################################################################### ### # # plugin-specific settings. # # any non empty hash set for a specific plugin will override the # general one above! # ################################################################### ### # # plugin eprints::plugin::import::marc::concordiatheses # $c->{marc}->{'eprints::plugin::import::marc::concordiatheses'}->{marc ep} = { ' a' => 'isbn', ' z' => 'isbn', ' a' => 'issn', ' a' => 'edition', information technology and libraries | march ' a' => 'place_of_pub', ' b' => 'publisher', ' c' => 'date', ' a' => 'pages_aacr', ' a' => 'volume', ' a' => 'series', ' c' => 'volume', ' x' => 'issn', ' a' => 'abstract', ' a' => 'publication', }; $c->{marc}->{'eprints::plugin::import::marc::concordiatheses'}->{constants} = { # marc to eprints constants 'type' => 'thesis', 'institution' => 'concordia university', 'date_type' => 'submitted', }; batch ingesting into eprints digital repository software| neugebauer and han appendix b. attach_documents.pl #!/usr/bin/perl -i/opt/eprints /perl_lib =head description this script allows you to attach a file to an eprint object by proquest id. =head copyright and license adam field, tomasz neugebauer bin han this module is free software under the same terms of perl. compatible with eprints . . (victoria sponge). =cut use strict; use warnings; use eprints; my $repositoryid = 'library'; my $root_dir = '/opt/eprints /bin/import-data/proquest'; #location of pdf files my $dataset_id = 'inbox'; #change to 'eprint' if you want to run it over everything. my $depositor = 'batchimporter'; #limit import to $depositor’s inbox #global variables for log purposes my $int_live = ; #count of eprints moved to live archive with a document my $int_doc = ; #count of eprints that already have document attached my @array_doc; #ids of eprints that already have documents my $int_no_doc = ; #count of eprints moved to live with no document attached my @array_no_doc; #ids of eprints that have no documents my $int_no_proid = ; #count of eprints with no proquest id my @array_no_proid; #ids of eprints with no proquest id my $session = eprints::session->new( , $repositoryid); die "couldn't create session for $repositoryid\n" unless defined $session; #the hash contains all the files that need to be uploaded #the hash contains key-value pairs: (pq_id => filename) my $filemap = {}; load_filemap($root_dir); #get all eprints in inbox dataset my $dataset = $session->get_repository->get_dataset($dataset_id); #run attach_file on each eprint object $dataset->map($session, \&attach_file); information technology and libraries | march #output log for attachment print "#### $int_doc eprints already have document attached, skip ####\n @array_doc\n"; print "#### $int_no_proid eprints doesn't have proquest id, skip ####\n @array_no_proid\n"; print "#### $int_no_doc eprints doesn't have associated document, moved to live ####\n @array_no_doc\n"; #total number of eprints that were made live: those with and without documents. my $int_total_live = $int_live + $int_no_doc; print "#### intotal: $int_total_live eprints moved to live ####\n"; #attach file to corresponding eprint object sub attach_file { my ($session, $ds, $eprint) = @_; #skip if eprint already has a document attached my $full_text_status = $eprint->get_value( "full_text_status" ); if ($full_text_status ne "none") { print "eprint ".$eprint->get_id." already has a document, skipping\n"; $int_doc ++; push ( @array_doc, $eprint->get_id ); return; } #retrieve username/userid associated with current eprint my $user = new eprints::dataobj::user( $eprint->{ session }, $eprint->get_value( "userid" ) ); my $username; # exit in case of failure to retrieve associated user, just in case. return unless defined $user; $username = $user->get_value( "username" ); # $dataset includes all eprints in inbox, so we limit to $depositor's items only return if( $username ne $depositor ); #skip if no proquest id is associated with the current eprint my $pq_id = $eprint->get_value('pq_id'); if (not defined $pq_id) { print "eprint ".$eprint->get_id." doesn't have a proquest id, skipping\n"; $int_no_proid ++; batch ingesting into eprints digital repository software| neugebauer and han push ( @array_no_proid, $eprint->get_id ); return; } #remove space from proquest id $pq_id =~ s/\s//g; #attach the pdf to eprint objects and move to live archive if ($filemap->{$pq_id} and -e $filemap->{$pq_id} ) #if the file exists { #create document object, add pdf files to document, attach to eprint object, and move to live archive my $doc = eprints::dataobj::document::create( $session, $eprint ); $doc->add_file( $filemap->{$pq_id}, $pq_id . '.pdf' ); $doc->set_value( "format", "application/pdf" ); $doc->commit(); print "adding document to eprint ", $eprint->get_id, "\n"; $eprint->move_to_archive; print "eprint ".$eprint->get_id." moved to archive.\n"; $int_live ++; } else { #move the metadata-only eprints to live as well print "proquest id \\$pq_id\\ (eprint ", $eprint->get_id, ") does not have a file associated with it\n"; $eprint->move_to_archive; print "eprint ".$eprint->get_id." moved to archive without document attached.\n"; $int_no_doc ++; push ( @array_no_doc, $eprint->get_id ); } } #recursively traverse the directory, find all pdf files. sub load_filemap { my ($directory) = @_; foreach my $filename (<$directory/*>) { if (-d $filename) { load_filemap($filename); } #catch the file name ending in .pdf elsif ($filename =~ m/([^\/]*)\.pdf$/i) information technology and libraries | march { my $pq_id = $ ; #add pq_id => filename pair to filemap hash table $filemap->{$pq_id} = $filename; } } } investigations into library web-scale discovery services jason vaughan information technology and libraries | march abstract web-scale discovery services for libraries provide deep discovery to a library’s local and licensed content and represent an evolution—perhaps a revolution—for end-user information discovery as pertains to library collections. this article frames the topic of web-scale discovery and begins by illuminating web-scale discovery from an academic library’s perspective—that is, the internal perspective seeking widespread staff participation in the discovery conversation. this included the creation of the discovery task force, a group that educated library staff, conducted internal staff surveys, and gathered observations from early adopters. the article next addresses the substantial research conducted with library vendors that have developed these services. such work included drafting of multiple comprehensive question lists distributed to the vendors, onsite vendor visits, and continual tracking of service enhancements. together, feedback gained from library staff, insights arrived at by the discovery task force, and information gathered from vendors collectively informed the recommendation of a service for the unlv libraries. introduction web-scale discovery services, combining vast repositories of content with accessible, intuitive interfaces, hold the potential to greatly facilitate the research process. while the technologies underlying such services are not new, commercial vendors releasing such services, and their work and agreements with publishers and aggregators to preindex content, is very new. this article in particular frames the topic of web-scale discovery and helps illuminate some of the concerns and commendations related to web-scale discovery from one library’s staff perspective—that is, the internal perspective. the second part focuses on detailed dialog with the commercial vendors, enabling the library to gain a better understanding of these services. in this sense, the second half is focused externally. given that web-scale discovery is new for the library environment, the author was unable to find any substantive published work detailing identification, research, evaluation, and recommendation related to library web-scale discovery services. it’s hoped that this article will serve as the ideal primer for other libraries exploring or contemplating exploration of these groundbreaking services. web-scale discovery services are able to index a variety of content, whether hosted locally or remotely. such content can include library ils records, digital collections, institutional repository content, and content from locally developed and hosted databases. such capabilities existed, to varying degrees, in next-generation library catalogs that debuted in the mid s. in addition, web-scale discovery services pre–index remotely hosted content, whether purchased or licensed by the library. this latter set of content—hundreds of millions of items—can include items such as e-books, publisher or aggregator content for tens of thousands of full-text journals, content from abstracting and indexing databases, and materials housed in open-access repositories. for purposes of this article, web-scale discovery services are flexible services which jason vaughan (jason.vaughan@unlv.edu) is director, library technologies, university of nevada, las vegas. investigations into library web-scale discovery services | vaughan provide quick and seamless discovery, delivery, and relevancy-ranking capabilities across a huge repository of content. commercial web-scale discovery vendors have brokered agreements with content providers (publishers and aggregators), allowing them to pre–index item metadata and full-text content (unlike the traditional federated search model). this approach lends itself to extremely rapid search and return of results ranked by relevancy, which can then be sorted in various ways according to the researcher’s whim (publication date, item type, full text only, etc.). by default, an intuitive, simple, google-like search box is provided (along with advanced search capabilities for those wishing this approach). the interface includes design cues expected by today’s researchers (such as faceted browsing) and, for libraries wishing to extend and customize the service, embraces an open architecture in comparison to traditional ils systems. why web-scale discovery? as illustrated by research dating back primarily to the s, library discovery systems within the networked online environment have evolved, yet continue to struggle to serve users. as a result, the library (or systems supported and maintained by the library) is often not the first stop for research—or worse, not a stop at all. users accustomed to a quick, easy, “must have it now” environment have defected, and research continues to illustrate this fact. rather than weave these research findings into a paragraph or page, below are some illustrative quotes to convey this challenge. the quotations below were chosen because they succinctly capture findings from research involving dozens, hundreds, and in some cases thousands of participants or respondents: people do not just use information that is easy to find; they even use information that they know to be of poor quality and less reliable—so long as it requires little effort to find—rather than using information they know to be of high quality and reliable, though harder to find. * * * today, there are numerous alternative avenues for discovery, and libraries are challenged to determine what role they should appropriately play. basic scholarly information use practices have shifted rapidly in recent years, and as a result the academic library is increasingly being disintermediated from the discovery process, risking irrelevance in one of its core functional areas [that of the library serving as a starting point or gateway for locating research information] . . . we have seen faculty members steadily shifting towards reliance on network- level electronic resources, and a corresponding decline in interest in using locally provided tools for discovery. * * * a seamless, easy flow from discovery through delivery is critical to end users. this point may seem obvious, but it is important to remember that for many end users, without the delivery of something he or she wants or needs, discovery alone is a waste of time. * * * end users’ expectations of data quality arise largely from their experiences of how information is organized on popular web sites. . . * * * [user] expectations are increasingly driven by their experiences with search engines like google and online bookstores like amazon. when end users conduct a search in a library information technology and libraries | march catalog, they expect their searches to find materials on exactly what they are looking for; they want relevant results. * * * users don’t understand the difference in scope between the catalog and a&i services (or the catalog, databases, digitized collections, and free scholarly content). * * * it is our responsibility to assist our users in finding what they need without demanding that they acquire specialized knowledge or select among an array of “silo” systems whose distinctions seem arbitrary . . . the continuing proliferation of formats, tools, services, and technologies has upended how we arrange, retrieve, and present our holdings. our users expect simplicity and immediate reward and amazon, google, and itunes are the standards against which we are judged. our current systems pale beside them. * * * q: if you could provide one piece of advice to your library, what would it be? a: just remember that students are less informed about the resources of the library than ever before because they are competing heavily with the internet. additional factors sell the idea of web-scale discovery. obviously, something must be discoverable for it to be used (and of value) to a researcher; ideally, content should be easily discoverable. since these new services index content that previously was housed in dozens or hundreds of individual silos, they can greatly facilitate the search process for many research purposes. libraries often spend large sums of money to license and purchase content, sums that often increase annually. any tool that holds the potential to significantly increase the discovery and use of such content should cause libraries to take notice. at time of writing, early research is beginning to indicate that these tools can increase discovery. doug way compared link-resolver-database and full-text statistics prior to and after grand valley state university’s implementation of the summon web- scale discovery service. his research suggests that the service was both broadly adopted by the university’s community and that it has led to an increase in their library’s electronic resource discovery and use. willamette university implemented worldcat local, and bill kelm presented results that showed an increase in both ill requests as well as use of the library’s electronic resources. from another angle, information-literacy efforts focus on connecting users to “legitimate” content and providing researchers the skills to identify content quality and legitimacy. given that these web-scale discovery services include or even primarily focus on indexing a large amount of scholarly research, such services can serve as another tool in the library’s arsenal. results retrieved from these services—largely content licensed or purchased by libraries—is accurate, relevant, and vetted, compared to the questionable or opinionated content that may often be returned through a web search engine query. several of the services currently allow a user to refine results to just categorized as peer-reviewed or scholarly. the internal academic library perspective: genesis of the unlv libraries discovery task force the following sections of this article begin with a focus on the internal unlv library perspective—from early discussions focused on the broad topic of discovery to establishing a task investigations into library web-scale discovery services | vaughan force charged to identify, research, evaluate, and recommend a potential service for purchase. throughout this process, and as detailed below, communication with and feedback from the variety of library staff was essential in ensuring success. given the increasing vitality of content in electronic format, and the fact that such content was increasingly spread across multiple access points or discovery systems, in late the university of nevada las vegas (unlv) libraries began an effort to engage library staff in information discovery and how such discovery would ideally occur in the future. related to the exponential growth of content in electronic format, traditional technical-services functions of cataloging and acquisitions were changing or would soon change, not just at unlv, but throughout the academic library community. coinciding with this, the libraries were working on drafting their – strategic plan and wanted to have a section highlighting the importance of information discovery and delivery with action items focused on improving this critical responsibility of libraries. in spring , library staff were given the opportunity to share with colleagues a product or idea, related to some aspect of discovery, which they felt was worthy of further consideration. this event, open to unlv libraries staff and other nevada colleagues, was titled the discovery mini-summit, and more than a dozen participants shared their ideas, most in a poster-session format. one of the posters focused on serial solutions summon, an early entrant into the vendor web-scale discovery service landscape. at the time, it was a few months from public release. other posters included topics such as the flickr commons (cultural heritage and academic institutions exposing their digital collections through this popular platform), and a working prototype of a homegrown, open-source federated search approach searching across various subscribed databases. in august , the dean of the unlv university libraries charged a ten-person task force to investigate and evaluate web-scale discovery services with the ultimate goal of providing a final recommendation for potential purchase. representation on the task force included three directors and a broad cross section of staff from across the functional areas of the library, including back-of-the-house and public-service operations. the director of library technologies, and author of this article, was tasked with drafting a charge and chairing the committee; once charged, the discovery task force worked over the next fifteen months to research, evaluate, and ultimately provide a recommendation regarding a web-scale discovery service. to help illustrate some of the events described, a graphical timeline of activities is presented as appendix a; the original charge appears as appendix b. in retrospect, the initial target date of early to make a recommendation was naive, as three of the five products ultimately identified and evaluated by the task force weren’t publicly released until . several boundaries were provided within the charge, including the fact that the task force was not investigating and evaluating traditional federated search products. the libraries had had a very poor experience with federated search a few years earlier, and the shortcomings of the traditional federated search approach—regardless of vendor—are well known. the remainder of this article discusses the various steps taken by the discovery task force in evaluating and researching web-scale discovery services. while many libraries have begun to implement the web- scale discovery services evaluated by this group, many more are currently at the learning and evaluation stage, or have not yet begun. many libraries that have already implemented a commercial service likely went through an evaluation process, but perhaps not at the scale conducted by the unlv libraries, if for no other reason than the majority of commercial services are extremely new. even in early , there was less competition, fewer services to evaluate, information technology and libraries | march fewer vendors to contact, and fewer early adopters from whom to seek references. fortunately, the initial target date of early for a recommendation was a soft target, and the discovery task force was given ample time to evaluate the products. based on presentations given by the author in , it can’t be presumed that an understanding of web-scale discovery—or the awareness of the commercial services now available—is necessarily widespread. in that sense, it’s the author’s hope and intent that information contained in this article can serve as a primer, or a recipe, for those libraries wishing to learn more about web-scale discovery and perhaps begin an evaluation process of their own. while research exists on federated search technologies within the library environment, the author was unable to find any peer-reviewed published research on the evaluation model and investigations for vendor produced web-scale discovery services as described in this paper. however, some reports are available on the open web, providing some insights into web-scale discovery evaluations led by other libraries, such as two reports provided by oregon state university. the first, dated march , describes a task force whose activities included “scrutinize wcl [worldcat local], investigate other vendors’ products, specifically serials solutions’ summon, the recently announced federated index discovery system; ebsco’s integrated search; and innovative interfaces’ encore product, so that a more detailed comparison can be done,” and “by march , communicate . . . whether wcl or another discovery service is the optimal purchase for osu libraries.” note that in , encore existed as a next-generation discovery layer, and it had an optional add on called “encore harvester,” which allows for the harvesting of digital local collections. the report cites the university of michigan’s evaluation of wcl, and adds their additional observations. the march report provides a features comparison matrix for worldcat local, encore, summon, and libraryfind (an open-source search tool developed at osu that provides federated searching for selected resources). feature sets include the areas of search and retrieval, content, and added features (e.g., book covers, user tagging, etc.). the report also describes some usability testing involving wcl and integration with other local library services. a second set of investigations followed “in order to provide the task force with an opportunity to more thoroughly investigate other products” and is described in a second report provided at the end of . at the time of both phases of this evaluation (and drafted reports) three of the web-scale discovery products had yet to enter public release. the december report focused on the two released products, serials solutions summon and worldcat local, and includes a feature matrix like the earlier report, with the added feature set of “other,” which included the features of “clarity of display,” “icons/images,” and “speed.” the latter report briefly describes how they obtained subject librarian feedback and the pros and cons observed by the librarians in looking at summon. it also mentions obtaining feedback from two early adopters of the summon product, as well as obtaining feedback from librarians whose library had implemented worldcat local. apart from the oregon reports, some other reports on evaluations (or selection) of a particular service, or a set of particular services, are available, such as the university of michigan’s article discovery working group, which submitted a final report in january . activity: understanding web-scale the first activity of the discovery task force was to educate the members, and later, other library colleagues, on web-scale discovery. terms such as “federated search,” “metasearch,” “next investigations into library web-scale discovery services | vaughan generation catalogs,” and “discovery layers” had all come before, and “web-scale” was a rather new concept that wasn’t widely understood. the discovery mini summit served as a springboard that perhaps more by chance than design introduced to unlv library staff what would later become more commonly known as web-scale discovery, though even we weren’t familiar with the term back in spring . in fall , the discovery task force identified reports from entities such as oclc, ithaka, and reports prepared for the library of congress highlighting changing user behavior and expectations; these reports helped form a solid foundation for understanding the “whys” related to web-scale discovery. additional registration and participation in sponsored web-scale discovery webcasts and meeting with vendors at library conferences helped further the understanding of web-scale discovery. after the discovery task force had a firm understanding of web-scale discovery, the group hosted a forum for all library staff to help explain the concept of web-scale discovery and the role of the discovery task force. specifically, this first forum outlined some key components of a web-scale discovery service, discussed research the task force had completed to date, and outlined some future research and evaluation steps. a summary of these steps appears in the timeline in appendix a. time was allowed for questions and answers, and then the task force broadcast several minutes of a (then recent) webcast talking about web-scale discovery. as part of its education role, the discovery task force set up an internal wiki-based webpage in august upon formation of the group, regularly added content, and notified staff when new content was added. a goal of the task force was to keep the evaluative process transparent, and over time the wiki became quite substantial. links to “live” services were provided on the wiki. given that some services had yet to be released, some links were to demo sites or sites of the closest approximation available, i.e., some services yet to be released were built on an existing discovery layer already in general release, and thus the look, feel, and functionality of such services was basically available for staff review. the wiki also provided links to published research and webcasts on web-scale discovery. such content grew over time as additional web- scale discovery products entered general release. in addition to materials on particular services, links were provided to important background documents and reports on topics related to the user discovery experience and user expectations for search, discovery, and delivery. discovery task force meeting notes and staff survey results were posted to the wiki, as were evaluative materials such as information on the content-overlap analysis conducted for each service. announcements to relevant vendor programs at the american library association’s annual conference were also posted to the wiki. activity: initial staff survey as noted above, when the task force began its work, only two products (out of five ultimately evaluated) were in general release. as more products entered public release, a next step was to invite vendors onsite to show their publicly released product, or a working, developed prototype nearing initial public release. to capture a sense of the library staff ahead of these vendor visits, the discovery task force conducted the first of two staff surveys. the -question survey consisted of a mix of “rank on a scale” questions, multiple-choice questions, and free-text response questions. both the initial and subsequent surveys were administered through the online surveymonkey tool. respondents were allowed to skip any question they wished. the survey was broken into three broad topical areas: “local library customization capabilities,” “end user aspect: information technology and libraries | march features and functionality,” and “content.” the survey had an average response rate of staff, or % of the library’s -strong workforce. the survey questions appear in appendix c. in hindsight, some of the questions could have benefitted from more careful construction. that said, there was a conscious juxtaposition of differing concepts within the same question—the task force did not want to receive a set of responses in which all library staff felt it was important for a service to do everything—in short, to be all things to all people. forcing staff to rate varied concepts within a question could provide insights into what they felt was really important. a brief summary of some key questions for each section follows. as an introduction, one question in the survey asked staff to rate the relative importance of each overarching aspect related to a discovery service (customization, end user interface, and content). staff felt content was the most critical aspect of a discovery service, followed by the end-user interface, followed by the ability to heavily customize the service. a snapshot of some of the capabilities library staff thought were important (or not) is provided in table . web-scale capabilities sa a n d sd physical item status information . % . % - - - publication date sort capability . % . % - - - display library-specified links in the interface . % . % - - - one-click retrieval of full-text items . % . % - - % ability to place ill / consortial catalog requests . % . % . % - - display the library’s logo . % . % . % - - to be embedded within various library website pages % % - - - full-text items first sort capability . % . % . % . % - shopping cart for batch printing, emailing, saving . % . % - - - faceted searching . % . % . % - - media type sort capability . % . % . % . % - author name sort capability . % . % . % . % - have a search algorithm that can be tweaked by library staff % % % % % user account for saved searches and marked items . % . % . % . % - book cover images % . % . % . % . % have a customizable color scheme % % % % - google books preview button for book items . % . % . % . % - tag cloud . % . % . % . % - user authored ratings . % . % . % . % . % user authored reviews . % . % % . % . % user authored tags . % . % . % . % . % sa = strongly agree; a = agree; n = neither agree nor disagree; d = disagree; sd = strongly disagree table . web-scale discovery service capabilities investigations into library web-scale discovery services | vaughan none of the results was surprising, other than perhaps the low interest or indifference in several web . community features, such as the ability for users to provide ratings, reviews, or tags for items, and even a tag cloud. the unlv libraries already had a next-generation catalog offering these features, and they have not been heavily used. even if there had been an appreciable adoption of these features by end users in the next-generation catalog for a web scale discovery service they are perhaps less applicable—it’s probably more likely that users would be less inclined to post reviews and ratings for an article, as opposed to a monograph—and article-level content vastly outnumbers book-level content with web-scale discovery services. the final survey section focused on content. one question asked about the incorporation of ten different information types (sources) and asked staff to rank how important it was that a service include such content. results are provided in table . a bit surprisingly, inclusion of catalog records was seen as most important. not surprisingly, full-text and a&i content from subscription resources were ranked very highly. it should also be noted that at the time of the survey, the institutional repository was in its infancy with only a few sample records, and awareness of this resource was low among library staff. another question listed a dozen existing publishers (e.g., springer, elsevier, etc.) deemed important to the libraries and asked staff to rank the importance that a discovery service index items from these publishers on a four point scale from “essential” to “not important.” results showed that all publishers were ranked as essential and important. related to content, . percent of staff felt that it was preferable for a service to de-dupe records such that the item appears once in the returned list of results; . percent preferred that the service not de-dupe results. information source rating average ils catalog records . majority of full-text articles / other research contained in vendor- licensed online resources . majority of citation records for non-full-text vendor-licensed a&i databases . consortial catalog records . electronic reserves records . records within locally created and hosted databases . digital collection records . worldcat records . ils authority control records . institutional repository records . table . importance of content indexed in discovery service after the first staff survey was concluded, the discovery task force hosted another library forum to introduce and “test drive” the five vendor services in front of library staff. this session was scheduled just a few weeks ahead of the onsite vendor visits to help serve as a primer to engage library staff and get them actively thinking about questions to ask the vendors. the task force information technology and libraries | march distributed notecards at the forum and asked attendees to record any specific questions they had about a particular service. after the forum, specific questions related to the particular products were collected; questions were collected, and they helped inform future research for those questions for which the task force did not at the time have an answer. questions ran the gamut and collectively touched on all three areas of evaluation. activity: second staff survey within a month after the five vendor onsite visits, a content analysis of the overlap between unlv licensed content and content indexed by the discovery services was conducted. after these steps, a second staff survey was administered. this second staff survey had questions focused on the same three functional areas as the first staff survey: local library customization features, end user features and functionality, and content. since the vendor visits had taken place and users could now understand the questions in the context of the products, questions were asked from the perspective of each product, e.g., “please rate on a five point likert scale whether each discovery service appears to adequately cover a majority of the critical publisher titles (worldcat local, summon, eds, encore synergy, primo central).” in addition, there were free-text questions focused on each individual product allowing colleagues to share additional, detailed thoughts. the second survey totalled questions and had an average response rate of respondents, or about percent of library staff. several staff conducted a series of sample searches in each of the services and provided feedback of their findings. though this was a small response rate, two of the five products rose to the top, a third was a strong contender, and two were seen as less desirable. the lower response rate is perhaps indicative of several things. first, not all staff had attended the onsite vendor demonstrations or had taken the time to test drive the services via the links provided on the discovery task force wiki site. second, some questions were more appropriately answered by a subset of staff. for example, the content questions might best be matched to those with reference, collection development, or curriculum and program liaison duties. finally, intricate details emerged once a thorough analysis of the vendor services was commenced. the first survey was focused more on the philosophy of what was desirable; the second survey took this a step further and asked how well each product matched such wishes. discovery services are changing rapidly with respect to interface updates, customization options, and scope of content. as such, and also reflective of the lower response rate, the author is not providing response information nor analysis for this second survey within this article. however, results may be provided upon specific request to the author. the questions themselves for the second staff survey are significant, and they could help serve as a model for other libraries evaluating existing services on the market. as such, questions appear in appendix d. activity: early adopter references one of the latter steps in the evaluation process from the internal academic library perspective was to obtain early adopter references from other academic library customers. a preliminary shortlist was compiled through a straw vote of the discovery task force—and the results of the vote showed a consensus. this vote narrowed down the discovery task force’s list of services still in contention for a potential purchase. this shortlist was based on the growing mass of research conducted by the discovery task force and informed by the staff surveys and feedback to date. three live customers were identified for each service that had made the shortlist, and the task investigations into library web-scale discovery services | vaughan force successfully obtained two references for each service. reference requests were intensive and involved a set of two dozen questions that references either responded to in writing or answered during scheduled conference calls. to help libraries conducting or interested in conducting their own evaluation and analysis of these services, this list of questions appears in appendix e. the services are so new that the live references weren’t able to comprehensively answer all the questions—they simply hadn’t had sufficient time to fully assess the service they’d chosen to implement. still, some important insights were gained about the specific products and, at the larger level, discovery services as a whole. as noted earlier, discovery services are changing rapidly in the sense of interface updates, customization options, and scope of content. as such, the author is not providing product specific response information or analysis of responses for each specific product—such investigations and interpretations are the job of each individual library seriously wishing to evaluate the services to help decide which product seems most appropriate for its particular environment. several broad insights merit notice, and they are shared below. regarding a question on implementation (though some challenges were mentioned with a few responders), nothing reached the threshold of serious concern. all respondents indicated the new discovery service is already the default or primary search box on their website. one section of the early adopter questions focused on content. the questions in this area seemed a bit challenging for the respondents to provide lots of detail. in terms of “adequately covering a majority of the important library titles,” respondents varied from “too early to tell,” “it covers many areas but there are some big names missing,” to two of the respondents answering simply, “yes.” several respondents also clearly indicated that the web-scale discovery service is not the “beginning and ending” for discovery, a fact that even some of the discovery vendors openly note. for example, one respondent indicated that web-scale discovery doesn’t replace remote federated searching. a majority (not all) of the discovery vendors also have a federated search product that can, to varying degrees, be integrated with their preharvested, centralized, index-based discovery service. this allows additional content to be searched because such databases may include content not indexed within the web-scale discovery service. however, many are familiar with the limitations of federated search technologies: slow speed, poor relevancy ranking of results, and the need to configure and maintain sources and targets. such problems remain with federated search products integrated with web-scale discovery services. another respondent indicated they were targeting their discovery service at undergraduate research needs. another responded, “as a general rule, i would say the discovery service does an excellent job covering all disciplines. if you start really in-depth research in a specific discipline, it starts to break down. general searches are great . . . dive deeper into any discipline and it falls apart. for example, for a computer science person, at some point they will want to go to acm or ieee directly for deep searches.” related to this, “the catalog is still important, if you want to do a very specific search for a book record, the catalog is better. the discovery service does not replace the catalog.” in terms of satisfaction with content type (newspapers, articles, proceedings, etc.), respondents seemed generally happy with the content mix. a range of responses were received, such as “doesn’t appear to be a leaning one way or another, it’s a mix. some of these things depend on how you set the system up, as there is quite a bit of flexibility; the library has to make a decision on what they want searched.” another example was that “the vendor has been working very hard to balance content types and i’ve seen a lot of improvement,” “no imbalance, results seem pretty well rounded.” another responded, “a common complaint is that newspapers and book reviews dominate the search results, but that is much more a function of search algorithms then the amount of content in the index.” information technology and libraries | march when asked about positive or critical faculty feedback to the service, several respondents indicated they hadn’t had a lot of feedback yet. one indicated they had anecdotal feedback. another indicated they’d received backlash from some users who were used to other search services (but also added that it was no greater than backlash from any other service they’d implemented in the past—and so the backlash wasn’t a surprise). one indicated “not a lot of feedback from faculty, the tendency is to go to databases directly, librarians need to instruct them in the discovery service.” for student feedback, one indicated, “we have received a few positive comments and see increased usage.” another indicated, “reviews are mixed. we have had a lot of feedback thanking us for providing a search that covers articles and books. they like the ability to do one search and get a mix of resources without the search taking a long time. other feedback usually centers around a bug or a feature not working as it should, or as they understand it should. in general, however, the feedback has been positive.” another replied, “comments we receive are generally positive, but we’ve not collected them systematically.” some respondents indicated they had done some initial usability testing on the initial interface, but not the most recent one now in use. others indicated they had not yet conducted usability testing, but it was planned for later in or . in terms of their fellow library staff and their initial satisfaction, one respondent indicated, “somewhere between satisfied and very satisfied . . . it has been increasing with each interface upgrade . . . our instruction librarians are not planning to use the discovery service this fall [in instruction efforts] because they need more experience with it . . . they have been overall intrigued and impressed by it . . . i would say our organization is grappling more with the implications of a discovery tools as a phenomenon than with our particular discovery service in particular. there seems to be general agreement that it is a good search tool for the unmediated searcher.” another indicated some concerns with the initial interface provided: “if librarians couldn’t figure it out, users can’t figure it out.” another responded, it was “a big struggle with librarians getting on board with the system and promoting the service to students. they continually compare it against the catalog. at one point, they weren’t even teaching the discovery service in bib instruction. the only way to improve things it with librarian feedback; it’s getting better, it has been hard. librarians have a hard time replacing the catalog and changing things that they are used to.” in terms of local customization, responses varied; some libraries had done basically no customization to the out-of-the-box interface, others had done extensive customization. one indicated they had tweaked sort options and added widgets to the interface. another indicated they had done extensive changes to the css. one indicated they had customized the colors, added a logo, tweaked the headers and footers, and created “canned” or preconfigured search boxes searching a subset of the index. another indicated they couldn’t customize the header and footer to the degree they would have liked, but were able to customize these elements to a degree. one respondent indicated they’d done a lot of customization to an earlier version of the interface, which had been rather painstaking, and that much of this broke when they upgraded to the latest version. that said, they also indicated the latest version was much better than the previous version. one respondent indicated it would be nice if the service could have multiple sources for investigations into library web-scale discovery services | vaughan enriched record content so that better coverage could be achieved. one respondent indicated they were working on a complete custom interface from scratch, which would be partially populated with results from the discovery service index (as well as other data sources). a few questions asked about relevancy as a search concept and how well the respondents felt about the quality of returned results for queries. one respondent indicated, “we have been able to tweak the ranking and are satisfied at this point.” another indicated, “overall, the relevance is good – and it has improved a lot.” another noted, “known item title searching has been a problem . . . the issues here are very predictable – one word titles are more likely to be a problem, as well as titles with stopwords,” and noted the vendor was aware of the issue and was improving this. one noted, “we would like to be able to experiment with the discovery service more – and noted, “no relevancy algorithm control.” another indicated they looked to investigate relevance more once usability studies commenced, and noted they worked with the vendor to do some code changes with the default search mechanism. one noted that they’d like to be able to specify some additional fields that would be part of the algorithm associated with relevancy. another optimistically noted “as an early adopter, it has been amazing to see how relevance has improved. it is not perfect, but it is constantly evolving and improving.” a final question asked simply, “overall, do you feel your selection of this vendor’s product was a good one? do you sense that your users – students and faculty – have positively received the product?” for the majority of responses, there was general agreement from the early adopters that they felt they’d made the right choice. one noted that it was still early and the evaluation is still a work in progress, but felt it has been positively received. the majority were more certain, “yes, i strongly feel that this was the right decision . . . as more users find it, i believe we will receive additional positive feedback,” “yes, we strongly believe in this product and feel it has been adopted and widely accepted by our users,” “i do feel it was a good selection.” the external perspective: dialog with web-scale discovery vendors the preceding sections focused on an academic library’s perspective on web-scale discovery services—the thoughts, opinions, preferences, and vetting activities involving library staff. the following sections focus on the extensive dialog and interaction with the vendors themselves, regardless of the internal library perspective, and highlight the thorough, meticulous research activities conducted on five vendor services. the discovery task force sought to learn as much about the each service as possible, a challenging proposition given the fact that at the start of investigations, only two of five services had been released, and, unsurprisingly, very little research existed. as such, it was critical to work with vendors to best understand their services, and how their service compared to others in the marketplace. broadly summarized efforts included identification of services, drafting of multiple comprehensive question lists distributed to the vendors, onsite vendor visits, and continual tracking of service enhancements. activity: vendor identification over the course of a year’s work, the discovery task force executed several steps to systematically understand the vendor marketplace—the capabilities, content considerations, development cycles, and future roadmaps associated with five vendor offerings. given that the information technology and libraries | march task force began their work when only two of these services were in public release, there was no manual, recipe, or substantial published research to rely on. the beginning, for the unlv libraries, lie in identification of the services—one must first know the services to be evaluated before evaluation can commence. as mentioned previously, the discovery mini-summit held at the unlv libraries highlighted one product—serial solutions summon; the only released product at the time of the mini-summit was worldcat local. while no published peer-reviewed research highlighting these new web-scale discovery services existed, press and news releases did exist for the three to-be-released services. such releases shed light on the landscape of services that the task force would review—a total of five services, from the first-to-market, worldcat local, to the most recent entrant, primo central. oclc worldcat local, released in november , can be considered the first web-scale discovery service as defined in this research; the experience of an early pilot partner (the university of washington) is profiled in a issue of library technology reports. in the uw pilot, approximately million article-level items were included with the worldcat database. another product, serials solutions summon, was released in july , and together these two services were the only ones publicly released when the discovery task force began its work. the task force identified three additional vendors each working on their own version of a web-scale discovery service; each of these services would enter initial general release as the task force continued its research: ebsco eds in january , innovative interfaces encore synergy around may , and ex libris primo central in june . while each of these three were new in terms of web-scale discovery capabilities, each was built, at least in part, on earlier systems from the vendors. eds draws heavily from the ebscohost interface (the original version of which dates back to the s), while the base encore and base primo systems were next-generation catalog systems that debuted in . activity: vendor investigations after identification of existing and under development discovery services, a next step in unlv’s detailed vendor investigations included the creation of a uniform, comprehensive question list sent to each of the five vendors. the discovery task force ultimately developed a list of questions divided into nine functional areas, as follows, with an example question: section : background. “when did product development begin (month, year)?” section : locally hosted systems and associated metadata. “with what metadata schemas does your discovery platform work? (e.g., marc, dublin core, ead, etc.)” section : publisher/aggregator coverage (full text and citation content). “with approximately how many publishers/aggregators have you forged content agreements ?” section : records maintenance and rights management. “how is your system initialized with the correct set of rights management information when a new library customer subscribes to your product?” investigations into library web-scale discovery services | vaughan section : seamlessness & interoperability with existing content repositories. “for ils records related to physical holdings, is status information provided directly within the discovery service results list?” section : usability philosophy. “describe how your product incorporates published, established best practices in terms of a customer focused, usable interface.” section : local “look & feel” customization options. “which of the following can the library control: color scheme; logo / branding; facet categories and placement; etc.” section : user experience (presentation, search functionality, and what the user can do with the results). “at what point does a user leave the context and confines of the discovery interface and enter the interface of a different system, whether remote or local?” section : administration module & statistics. “describe in detail the statistics reporting capabilities offered by your system. does your system provide the following sets of statistics . . .” all vendors were given – weeks to respond, and all vendors responded. it was evident from the uneven level of responses to the questions that the vendors were at different developmental states with their products. some vendors were still – months away from initial public release; some were not even firm on when their service would enter release. it was also observed that some vendors were less explicit in the level of detail provided, reflective of, or in some cases perhaps regardless of, development state. a refined subset of the original questions appears as a list of questions in appendix f. apart from the detailed question list, various sets of free and licensed information on these discovery services are available online, and the task force sought to identify and digest the information. the charleston advisor has conducted interviews with several of the library web- scale discovery vendors on their products, including ebsco, serials solutions, and ex libris. these interviews, each around a dozen questions, ask the vendors to describe their product, how it differs from other products in the marketplace, and include questions on metadata and content—all important questions. an article by ronda rowe reviews summon, eds, and worldcat local, and provides some analysis of each product on the basis of content, user interface and searchability, pricing, and contract options. it also provides a comparison of product features provided by these three services, such as “search box can be embedded in any webpage,” “local branding possible,” and “supports social networking.” a wide variety of archived webcasts, many provided by library journal, are available through free registration, and new webcasts are being offered at time of writing; these presentations to some degree touch on discussions with the discovery vendors, and are often moderated or include company representatives as part of the discussion group. several libraries have authored reports and presentations that, at least partially, discuss information on particular services gained through their evaluations, which include dialog with the vendors. vendors themselves each have a section on their corporate website devoted to their service. information provided on these websites ranges from extremely brief to, in the case of worldcat local, very detailed and informative. in addition, much can be gained by “test-driving” live implementations. as such, a listing of vendor website addresses information technology and libraries | march providing more information as well as a list of sample, live implementations is provided in appendix g. activities: vendor visits and content overlap analysis each of the five vendors visited the unlv libraries in spring . vendor visits all occurred within a nine-day span; visits were intentionally scheduled close to each other to keep things fresh in the minds of library staff, and such proximity would help with product comparisons. vendor visits lasted approximately half a day, and each vendor visit often included the field or regional sales representative as well as a product manager or technical expert. vendor visits included a demonstration and q&a for all library staff as well as invited colleagues from other southern nevada libraries, a meeting with the discovery task force, and a meeting with technical staff at unlv responsible for website design and application development and customization. vendors were each given a uniform set of fourteen questions on topics to address during their visit; these appear in appendix h. questions were divided into the broad topical areas of content coverage, end user interface and functionality, and staff “control” over the end user interface. on average, approximately – percent of the library staff attended the open vendor demo and q & a session. shortly after the vendor visits, a content-overlap analysis comparing unlv serials holdings with indexed content in the discovery service was sought from each vendor. given that the amount of content indexed by each discovery service was growing (and continues to grow) extremely rapidly as new publisher and aggregator content agreements are signed, this content-overlap analysis was intentionally not sought at an earlier date. some vendors were able to provide detailed coverage information against our existing journal titles (unlv currently subscribes to approximately , e-journals and provides access to another , + open-access titles). for others, this was more difficult. recognizing this, the head of collection development was asked to provide a list of the “top ” journal titles for unlv based on such factors as usage statistics and whether the title was a core title for part of the unlv curriculum. the remaining vendors were able to provide content coverage information against this critical title list. four of the five products had quite comprehensive coverage (more than percent) of the unlv libraries’ titles. while outside the scope of this article, “coverage” can mean different things for different services. driven by the publisher agreements they are able to secure, some discovery services may have extensive coverage for particular titles (such as the full text, abstracts, author-supplied keywords, subject headings, etc.), whereas other services, while covering the same title, may have “thinner” metadata, such as basic citation information (article title, publication title, author, publication date, etc.). more discussion on this topic is present in the january library technology reports on library web-scale discovery services. activity: product development tracking one aspect of web-scale discovery services, and the next-generation discovery layers that preceded them, is a rapid enhancement cycle, especially when juxtaposed against the turnkey- style ils system that dominated library automation for many years. as an example, minor enhancements are provided by serials solutions to summon approximately every three to four weeks; provided by ebsco to ebsco discovery service approximately every three months; and investigations into library web-scale discovery services | vaughan provided by ex libris to primo/primo central approximately every three months. many vendors unveil updates coinciding with annual library conferences, and was no exception. in late summer/early fall , the discovery task force had conference calls or onsite visits with several of the vendors with a focused discussion on new enhancements and changes to services as well as to obtain answers to any questions that arose since their last visit several months earlier. since the vendor visits in spring , each service had changed, and two services had unveiled significantly different and improved interfaces. the discovery task force’s understanding of web-scale discovery services had expanded greatly since starting their work. coordinated with the second series of vendor visits and discussions, an additional list of more than two dozen questions, recognizing this refined understanding, was sent to the majority of vendors. a portion of these questions are provided as part of the refined list of questions presented in appendix f. this second set of questions dealt with complex discussions of metadata quality, such as what level of content publishers and aggregators were providing for indexing purposes, e.g., full text, abstracts, table of contents, author-supplied keywords or subject headings, or particular citation and record fields), and also the vendor’s stance on content neutrality, i.e., whether they are entering into exclusive agreements with publishers and aggregators, and, if the discovery service vendor is owned by a company involved with content, if that content is promoted or weighted more heavily in result sets. other questions dealt with such topics as current install base counts and technical clarifications about how their service worked. in particular, the questions related to content were tricky for many (not all) of the vendors to address. still, the discovery task force was able to get a better understanding of how things worked in the evolving discovery environment. combined with the internal library perspective and the early adopter references, information gathered from vendors provided the necessary data set to submit a recommendation with confidence. activity: recommendation by mid-fall , the discovery task force had conducted and had at their disposal a tremendous amount of research. recognizing how quickly these services change and the fact that a cyclical evaluation could occur, the task force members felt they had met their charge. if all things failed during the next phase—implementation—at least no one would be able to question the thoroughness of the task force’s efforts. unlike the hasty decision, which in part led to a less than stellar experience with federated search a few years earlier, the evaluation process to recommend a new web-scale discovery service was deliberate, thorough, transparent, and vetted with library stakeholders. given the discovery task force was entering its final phase, official price quotes were sought from each vendor. each task force member was asked to develop a pro/con list for all five identified products based on the knowledge that was gained. these lists were anonymized and consolidated into a single, extensive pro/con list for each service. some of the pros and cons were subjective (such as the interface aesthetics), some were objective (such as a particular discovery service not offering a desired feature). at one of the final meetings of the task force, members reaffirmed the three top contenders, indicated the other two were no longer under consideration and, afterward, were asked to rank their first, second, and third choices for the remaining services. while complete consensus wasn’t achieved, there was a resounding first choice, second choice, and third information technology and libraries | march choice. the task force presented a summary of findings at a meeting open to all library staff. this meeting summarized the research and evaluation steps the task force had conducted over the past year, framed each of the three shortlisted services by discussing some strengths and weaknesses of each service as observed by the task force, and sought to answer any questions from the library at large. prior to drafting the final report and making the recommendation to the dean of libraries, several task force members led a discussion and final question and answer at a libraries’ cabinet meeting, one of the high-level administrative groups at the unlv libraries. vetting by this body represented the last step related to the discovery task force’s investigation, evaluation, and recommendation for purchase of a library web-scale discovery service. the recommendation was broadly accepted by the library cabinet, and shortly afterward the discovery task force was officially disbanded, having met its goal of investigating, evaluating, and making a recommendation for purchase of a library web-scale discovery service. next steps the dialog above describes the research, evaluation, and recommendation model used by the unlv libraries to select a web-scale discovery service. such a model and the associated appendixes could serve as a framework, with some adaptations perhaps, for other libraries considering the evaluation and purchase of a web-scale discovery service. together, the discovery task force’s internal and external research and evaluation provided a substantive base of knowledge on which to make a recommendation. after its recommendation, the project progressed from a research and recommendation phase to an implementation phase. the libraries’ cabinet brainstormed a list of more than a dozen concise implementation bullet points—steps that would need to be addressed—including the harvesting and metadata mapping of local library resources, local branding and some level of customization work, and integration of the web-scale discovery search box in the appropriate locations on the libraries’ website. project implementation co-managers were assigned (the director of technical services and the web technical support manager), as well as key library personnel who would aid in one or more implementation steps. in january , the implementation commenced, with an expected public launch of the new service planned for mid- . the success of a web-scale discovery service at the unlv libraries is a story yet to be written, but one full of promise. acknowledgements the author wishes to thank the other members of the unlv libraries’ discovery task force in the research and evaluation of library web-scale discovery services: darcy del bosque, alex dolski, tamera hanken, cory lampert, peter michel, vicki nozero, kathy rankin, michael yunkin, and anne zald. references . marcia j. bates, improving user access to library catalog and portal information, final report, version (washington, dc: library of congress, ), , http://www.loc.gov/catdir/bibcontrol/ . batesreport - .doc.pdf (accessed september , ). http://www.loc.gov/catdir/bibcontrol/ . batesreport - .doc.pdf http://www.loc.gov/catdir/bibcontrol/ . batesreport - .doc.pdf investigations into library web-scale discovery services | vaughan . roger c. schonfeld and ross housewright, faculty survey : key strategic insights for libraries, publishers, and societies (new york: ithaka s+r, ), , http://www.ithaka.org/ithaka-s-r/research/faculty-surveys- - /faculty% study% .pdf (accessed september , ). . oclc, online catalogs: what users and librarians want (dublin, oh: oclc, ), , http://www.oclc.org/reports/onlinecatalogs/fullreport.pdf (accessed september , ). . ibid, vi. . ibid, . . karen calhoun, the changing nature of the catalog and its integration with other discovery tools: final report (washington, dc: library of congress, ), , http://www.loc.gov/catdir/calhoun-report-final.pdf (accessed september , ). . bibliographic services task force, rethinking how we provide bibliographic services for the university of california: final report ([pub location?] university of california libraries, ), , http://libraries.universityofcalifornia.edu/sopag/bstf/final.pdf (accessed september , ). . oclc, college students’ perceptions of libraries and information resources (dublin, oh: oclc, ), part , page , http://www.oclc.org/reports/pdfs/studentperceptions.pdf (accessed september , ). . doug way, “the impact of web-scale discovery on the use of a library collection,” serials review, in press. . bill kelm, “worldcat local effects at willamette university,” presentation, prezi, july , , http://prezi.com/u pzunpb fa/worldcat-local-effects-at-wu/ (accessed sept , ). . michael boock, faye chadwell, and terry reese, “worldcat local task force report to lamp,”march , , http://hdl.handle.net/ / (accessed february , ). . michael boock et al., “discovery services task force recommendation to university librarian,” http://hdl.handle.net/ / (accessed february , ). . ken varnum et al., “university of michigan library article discovery working group final report,” umich, january , , http://www.lib.umich.edu/files/adwg/final-report.pdf.[access date?] . jennifer ward, pam mofjeld, and steve shadle, “worldcat local at the university of washington libraries,” library technology reports , no. (august/september ). . dennis brunning and george machovec, “an interview with sam brooks and michael gorrell on the ebscohost integrated search and ebsco discovery service,” charleston advisor , no. (january ): – . http://www.ithaka.org/ithaka-s-r/research/faculty-surveys- - /faculty% study% .pdf http://www.ithaka.org/ithaka-s-r/research/faculty-surveys- - /faculty% study% .pdf http://www.oclc.org/reports/onlinecatalogs/fullreport.pdf http://www.loc.gov/catdir/calhoun-report-final.pdf http://libraries.universityofcalifornia.edu/sopag/bstf/final.pdf http://www.oclc.org/reports/pdfs/studentperceptions.pdf http://prezi.com/u pzunpb fa/worldcat-local-effects-at-wu/ http://hdl.handle.net/ / http://hdl.handle.net/ / http://www.lib.umich.edu/files/adwg/final-report.pdf information technology and libraries | march . dennis brunning and george machovec, “interview about summon with jane burke, vice president of serials solutions,” charleston advisor , no. (april ): – . . dennis brunning and george machovec, “an interview with nancy dushkin, vp discovery and delivery solutions at ex libris, regarding primo central,” charleston advisor , no. (october ): – . . ronda rowe, “web-scale discovery: a review of summon, ebsco discovery service, and worldcat local,” charleston advisor , no. (october ): – . . library journal archived webcasts are available at http://www.libraryjournal.com/csp/cms/sites/lj/tools/webcast/index.csp (accessed sept , ). . boock, chadwell, and reese, “worldcat local task force report to lamp”; boock et al., “discovery services task force recommendation to university librarian”; ken varnum et al., “university of michigan library article discovery working group final report.” . jason vaughan, “library web-scale discovery services,” library technology reports , no. (january ). note: appendices a–h available as supplemental files. http://www.libraryjournal.com/csp/cms/sites/lj/tools/webcast/index.csp investigations into library web-scale discovery services: appendices a-h jason vaughan information technology and libraries | march appendices appendix a. discovery task force timeline appendix b. discovery task force charge appendix c. discovery task force: staff survey questions appendix d. discovery task force: staff survey questions appendix e. discovery task force: early adopter questions appendix f. discovery task force: initial vendor investigation questions appendix g. vendor websites and example implementations appendix h. vendor visit questions investigations into library web-scale discovery services | vaughan appendix a. discovery task force timeline information technology and libraries | march appendix b. discovery task force charge discovery task force charge informed through various efforts and research at the local and broader levels, and as expressed in the libraries / strategic plan, the unlv libraries have the desire to enable and maximize the discovery of library resources for our patrons. specifically, the unlv libraries seek a unified solution which ideally could meet these guiding principles: • creates a unified search interface for users pulling together information from the library catalog as well as other resources (e.g. journal articles, images, archival materials). • enhances discoverability of as broad a spectrum of library resources as possible • intuitive: minimizes the skills, time, and effort needed by our users to discover resources •supports a high level of local customization (such as accommodation of branding and usability considerations) • supports a high level of interoperability (easily connecting and exchanging data with other systems that are part of our information infrastructure) •demonstrates commitment to sustainability and future enhancements •informed by preferred starting points as such, the discovery task force advises libraries administration on a solution that appears to best meet the goal of enabling and maximizing the discovery of library resources. a bulk of the work will entail a marketplace survey and evaluation of vendor offerings. charge specific deliverables for this work include: . identify vendor next generation discovery platforms, whether established and currently on the market, or those publicized and at an advanced stage of development, with an expectation of availability within a year’s time. identify & create a representative list of other academic libraries which have implemented or purchased currently available products. . create a checklist / criteria of functional requirements / desires for a next generation discovery platform. . create lists of questions to distribute to potential vendors and existing customers of next generation discovery platforms. questions will focus on broad categories such as the following: a. seek to understand how content hosted in our current online systems (iii catalog, contentdm, locally created databases, vendor databases, etc.) could/would (or not be able investigations into library web-scale discovery services | vaughan to) be incorporated or searchable within the discovery platform. apart from our existing online systems as we know them today, the task force will explore, in general terms, how new information resources could be incorporated into the discovery platform. more explicitly, the task force will seek an understanding of what types of existing records are discoverable within the vendor’s next generation discovery platform, and seek an understanding of what basic metadata must exist for an item to be discoverable. b. seek to understand whether the solution relies on federated search, the creation of a central site index via metadata harvesting, or both, to enable discovery of items. c. additional questions, such as pricing, maintenance, install base, etc. . evaluate gathered information and seek feedback from library staff. . provide to the dean’s directs a final report which summarizes the task force findings. this report will include a recommended product(s) and a broad, as opposed to detailed, summary of workload implications related to implementation and ongoing maintenance. the final report should be provided to the dean’s directs by february , . boundaries the work of the task force does not include: • detailing the contents of “hidden collections” within the libraries and seeking to make a concrete determination that such hidden collections, in their current form, would be discoverable via the new system. • conducting an inventory, recommending, or prioritizing collections or items which should be cataloged or otherwise enriched with metadata to make them discoverable. • coordination with other southern nevada nshe entities. • an ils marketplace survey. the underlying innovative millennium system is not being reviewed for potential replacement. • implementation of a selected product. [the charge concluded with a list of members for the task force] information technology and libraries | march appendix c. discovery task force: staff survey questions “rank” means the surveymonkey question will be set up such that each option can only be chosen once, and will be placed on a scale that corresponds to the number of choices overall. “rate” means there will be a point likert scale ranging from strongly disagree to strongly agree. section : customization. the “staff side” of the house . customization. it is important for the library to be able to control/tweak/influence the following design element [strongly disagree / disagree / neither agree or disagree / agree / strongly agree]  general color scheme  ability to include a unlv logo somewhere on the page.  ability to add other branding elements to the page.  ability to add one or more library specified links prominently in the interface (example: a link to the libraries’ home page)  able to customize the name of the product (meaning, the vendor’s name for the product doesn’t need to be used nor appear within the interface)  ability to embed the search box associated with the discovery platform elsewhere into the library website, such as the homepage (i.e. the user could start a search w/o having to directly go to the discovery platform . customization. are there any other design customization capabilities that are significantly important? please list, and please indicate if this is a high, low, or medium priority in terms of importance to you. (freetext box ) . search algorithms. it is important for the library to be able to change or tweak the platform’s native search algorithm to be able to promote desired items such that they appear higher in the returned list of [strongly disagree / disagree / neither agree or disagree / agree / strongly agree] [e.g. the library, at its option, could tweak one or more search algorithms to more heavily weight resources it wants to promote. for example, if a user searches for “hoover dam” the library could set a rule that would heavily weight and promote unlv digital collection images for hoover dam – those results would appear on the first page of results]. . statistics. the following statistic is important to have for the discovery platform [strongly disagree / disagree / neither agree or disagree / agree / strongly agree]  number of searches, by customizable timeframe number of item or article level records accessed (that is, a user clicks on something in the returned list of results)  number of searches generating results investigations into library web-scale discovery services | vaughan  number of items accessed by type  number of items accessed by provider of content (that is, number of articles from particular database/fulltext vendor . statistics. what other statistics would you like to see a discovery platform provide and how important is this to you? (freetext box) . staff summary. please rank on a - scale how important the following elements are, with a “ ” being most important, a “ ” being nd most important, and a being rd most important.  heavy customization capabilities as described in questions & above  ability to tweak search algorithms as described in question  ability for the system to natively provide detailed search stats such as described in question , . section . the “end user” side of the house . searching. which of the following search options is preferable when a user begins their search [choose one]  the system has a “google-like” simple search box  the system has a “google-like” simple search box, but also has an advanced search capability (user can refine the search to certain categories: author, journal, etc.)  no opinion . zero hit searches. for a search that retrieves no actual results: [choose one]  the system should suggest something else or ask, “did you mean?”  retrieving precise results is more important and the system should not suggest something else or ask “did you mean?”  no opinion . de-duplication of similar items. which of the following is preferable [choose one]  the system automatically de-dupes records (the item only appears once in the returned list)  the system does not de-dupe records (the same item could appear more than once in the returned list, such as when we have overlapping coverage of a particular journal from multiple subscription vendors)  no opinion information technology and libraries | march . sorting of returned results. it is important for the user to be able to sort or reorder a list of returned results by . . [strongly disagree / disagree / neither agree or disagree / agree / strongly agree]  publication date  alphabetical by author name  alphabetical by title  full text items first  by media type (examples: journal, book, image, etc) . web . functionality on returned results. the following items are important for a discovery platform to have . . [strongly disagree / disagree / neither agree or disagree / agree / strongly agree] (note, if necessary, please conduct a search in the libraries’ encore system to help illustrate / remember some of the features/jargon mentioned below. in encore, “facets” appear on the left hand side of the screen; the results with book covers, “add to cart,” and “export” features appear in the middle; and a tag cloud to the right. note: this question is asking about having the particular feature regardless of which vendor, and not how well or how poorly you think the feature works for the encore system)  a tag cloud  faceted searching  ability to add user-generated tags to materials (“folksonomies”)  ability for users to write and post a review of an item • other (please specify) . enriched record information on returned results. the following items are important to have in the discovery system . . . [strongly disagree / disagree / neither agree or disagree / agree / strongly agree]  book covers for items held by the libraries  a google books preview button for print items held by the libraries  displays item status information for print items held by the libraries (example: available, checked out) . what the user can do with the results. the following functionality is important to have in the discovery system . . [strongly disagree / disagree / neither agree or disagree / agree / strongly agree]  retrieve the fulltext of an item with only a single click on the item from the initial list of returned results  ability to add items to a cart for easy export (print, email, save, export to refworks) investigations into library web-scale discovery services | vaughan  ability to place an interlibrary loan / link+ request for an item  system has a login/user account feature which can store user search information for later. in other words, a user could potentially log in to retrieve saved searches, previously stored items, or create alerts when new materials become available. . miscellaneous. the following feature/attribute is important to have in the discovery system . . . [strongly disagree / disagree / neither agree or disagree / agree / strongly agree]  the vendor has an existing mobile version of their discovery tool for use by smartphones or other small internet-enabled devices.  the vendor has designed the product such that it can be incorporated into other sites used by students, such as webcampus and/or social networking sites. such “designs” may include the use of persistent urls to embed hyperlinks, the ability to place the search box in another website, or specifically designed widgets developed by the vendor  indexing and availability of newly published items occurs within a matter of days as opposed to a week or perhaps a month.  library catalog authority record information is used to help return proper results and/or populate a tag cloud. . end user summary. please rank on a - scale how important the following elements are; a “ ” means you think it is the most important, a “ ” second most important, etc.  system offers a “google-like” simple search box only, as detailed in question above  system offers a “did you mean?” or alternate suggestions for all searches retrieving results as detailed in question above (obviously, if you value precision of results over “did you mean” functionality, you would rank this toward the lower end of the spectrum).  system de-dupes similar items as detailed in question above(if you believe the system should not de- dupe similar items, you would rate this toward the lower end of the spectrum)  system provides multiple sort options of returned results as detailed in question above  system offers a variety of web . features as detailed in question above  system offer enriched record information as detailed in question above  system offers flexible options for what a user can do with the results, as detailed in question above  system has one or more miscellaneous features as detailed in question above. section : content . incorporation of different information types. in an ideal world, a discovery platform would incorporate all of our electronic resources, whether locally produced or licensed/purchased from vendors. below is a listing of different information types. please rank on a scale of - how vital it is information technology and libraries | march that a discovery platform accommodate these information types (“ ” is the most important item in your mind, a “ ” is second most important, etc). a. innopac millennium records for unlv print & electronic holdings b. link+ records for print holdings held within the link+ consortium c. innopac authority control records d. records within oclc worldcat e. contentdm records for digital collection materials f. bepress digital commons institutional repository materials g. locally created web accessible database records (e.g. the special collections & architecture databases) h. electronic reserves materials hosted in eres i. a majority of the citation records from non fulltext, vendor licensed online index/abstract/citation databases (e.g. the “agricola” database) j. a majority of the fulltext articles or other research contained in many of our vendor licensed online resources (e.g. “academic search premier” which contains a lot of full text content, and the other fulltext resource packages / journal titles we subscribe to) . local content. related to item (g) in the question immediately above, please list any locally produced collections that are currently available either on the website, or in electronic format as a word document, excel spreadsheet or access database (and not currently available on the website) that you would like the discovery platform to incorporate. (freetext box) . particular sets of licensed resources, what’s important? please rank which of the licensed (full text or primarily full text) existing publishers below are most important for a discovery platform to accommodate. elsevier sage wiley springer american chemical society taylor & francis (informaworld) ieee american institute of physics oxford ovid nature emerald investigations into library web-scale discovery services | vaughan section : survey summary . overarching survey question. the questions above were roughly categorized into three areas. given that no discovery platform will be everything to everybody, please rank on a - scale what the most important aspects of a discovery system are to you ( is most critical, is second in importance overall, etc.)  the platform is highly customizable by staff (types of things in area of the survey)  the platform is highly flexible from the end-user standpoint (type of things in area of the survey)  the platform encompasses a large variety of our licensed and local resources (type of things in area of the survey) . additional input. the survey above is roughly drawn from a larger list of questions sent to the discovery task force vendors. what other things do you think are really important when thinking about a next-generation discovery platform? (freetext input, you may write a sentence or a book) . demographic. what library division do you belong to? library administration library technologies research & education special collections technical services user services information technology and libraries | march appendix d. discovery task force: staff survey question for the comparison questions, products are listed by order of vendor presentation. please mark an answer for each product. part i. licensed publisher content (e.g. fulltext journal articles; citations / abstracts) sa = strongly agree; a = agree; n= neither agree nor disagree; d = disagree; sd = strongly disagree . “the discovery platform appears to adequately cover a majority of the critical publisher titles.” sa a n d sd i don’t know enough about the content coverage for this product to comment ex libris primo central oclc worldcat local ebsco discovery services innovative encore synergy serials solutions summon . “the discovery platform appears to adequately cover a majority of the second-tier or somewhat less critical publisher titles.” sa a n d sd i don’t know enough about the content coverage for this product to comment ex libris primo central oclc worldcat local ebsco discovery services innovative encore synergy serials solutions summon . overall, from the content coverage point of view, please rank each platform from best to worst. worst nd worst middle nd best best ex libris primo central oclc worldcat local ebsco discovery services innovative encore synergy serials solutions summon . regardless of a best to worst ranking, please indicate if the products were, overall, acceptable or unacceptable to you from the content coverage standpoint. unacceptable acceptable ex libris primo central investigations into library web-scale discovery services | vaughan oclc worldcat local ebsco discovery services innovative encore synergy serials solutions summon part ii. end-user functionality & ease of use . from the user perspective, how functional do you think the discovery platform is? are the facets and/or other methods that one can use to limit or refine a search appropriate? were you satisfied with the export options offered by the system (email, export into refworks, print, etc.)? if you think web . technologies are important (tag cloud, etc.), were one or more of these present (and well executed) in this product? the platform appears to be severely limited in major aspects of end user functionality the platform appears to have some level of useful functionality, but perhaps not as much or as well executed as some competing products. yes, the platform seems quite rich in terms of end user functionality, and such functions are well executed. i can’t comment on this particular product because i didn’t see the vendor demo, haven’t visited any of the live implementations linked on the discovery wiki page, or otherwise don’t have enough information. ex libris primo central oclc worldcat local ebsco discovery services innovative encore synergy serials solutions summon . from the user perspective, for a full-text pdf journal article, how easy is it to retrieve the full-text? does it take many clicks? are there confusing choices? it’s very cumbersome trying to retrieve the full text of an item, there are many clicks, and/or it’s simply confusing when going through the steps to retrieve the full text. it’s somewhat straightforward to retrieve a full text item, but perhaps it’s not as easy or as well executed as some of the competing products it’s quite easy to retrieve a full text item using this platform, as good as or better than the competition, and i don’t feel it would be a barrier to a majority of our users. i can’t comment on this particular product because i didn’t see the vendor demo, haven’t visited any of the live implementations linked on the discovery wiki page, or otherwise don’t have enough information. ex libris primo central information technology and libraries | march oclc worldcat local ebsco discovery services innovative encore synergy serials solutions summon . how satisfied were you with the platform’s handling of “dead end” or “zero hit” searches? did the platform offer “did you mean” spelling suggestions? did the platform offer you the option to request the item via doc delivery / link+? is the vendor’s implementation of such features well executed, or were they difficult, confusing, or otherwise lacking? the platform appears to be severely limited in or otherwise poorly executes how it responds to a dead end or zero hit search. the platform handled dead end or zero hit results, but perhaps not as seamlessly or as well executed as some of the competing products. i was happy with how the platform handled “dead end” searches, and such functionality appears to be well executed, as good as or better than the competition. i can’t comment on this particular product because i didn’t see the vendor demo, haven’t visited any of the live implementations linked on the discovery wiki page, otherwise don’t have enough information. ex libris primo central oclc worldcat local ebsco discovery services innovative encore synergy serials solutions summon . how satisfied were you with the platform’s integration with the opac? were important things such as call numbers, item status information, and enriched content immediately available and easily viewable from within the discovery platform interface, or did it require an extra click or two into the opac – and did you find this cumbersome or confusing? the platform provides minimal opac item information, and a user the platform appeared to integrate ok with the opac in i was happy with how the platform integrated with the i can’t comment on this particular product because i didn’t see the investigations into library web-scale discovery services | vaughan would have to click through to the opac to get the information they might really need; and/or it took multiple clicks or was otherwise cumbersome to get the relevant item level information terms of providing some level of relevant item level information, but perhaps not as much or as well executed as competing products. opac. a majority of the opac information was available in the discovery platform, and/or their connection to the opac was quite elegant. vendor demo, haven’t visited any of the live implementations linked on the discovery wiki page, or otherwise don’t have enough information. ex libris primo central oclc worldcat local ebsco discovery services innovative encore synergy serials solutions summon . overall, from an end user functionality / ease of use standpoint – how a user can refine a search, export results, easily retrieve the fulltext, easily see information from the opac record – please rank each platform from best to worst. worst nd worst middle nd best best ex libris primo central oclc worldcat local ebsco discovery services innovative encore synergy serials solutions summon . regardless of a best to worst ranking, please indicate if the products were, overall, acceptable or unacceptable to you from the user functionality / ease of use standpoint. unacceptable acceptable ex libris primo central oclc worldcat local ebsco discovery services innovative encore synergy serials solutions summon part iii. staff customization information technology and libraries | march . the “out of the box” design demo’ed at the presentation (or linked to the discovery wiki page – whichever particular implementation you liked best for that product) was . . seriously lacking and i feel would need major design changes and customization by library web technical staff. middle of the road – some things i liked, some things i didn’t. the interface design was better than some competing products, worse than others. appeared very professional, clean, well organized, and usable; the appearance was better than most/all of the others products. i can’t comment on this particular product because i didn’t see the vendor demo, haven’t visited any of the live implementations linked on the discovery wiki page, or otherwise don’t have enough information. ex libris primo central oclc worldcat local ebsco discovery services innovative encore synergy serials solutions summon . all products offer some level of customization options that allow at least some changes to the “out of the box” platform. based on what the vendors indicated about the level of customization possible with the platform (e.g. look and feel, ability to add library links, ability to embed the search box on a homepage) do you feel there is enough flexibility with this platform for our needs? the platform appears to be severely limited in the degree or types of customization that can occur at the local level. we appear “stuck” with what the vendor gives us – for better or worse. the platform appeared to have some level of customization, but perhaps not as much as some competing products. yes, the platform seems quite rich in terms of customization options under our local control; more so than the majority or all of the other products. i can’t comment on this particular product because i didn’t see the vendor demo, don’t have enough information, and/or would prefer to leave this question to technical staff to weigh in on. ex libris primo central oclc worldcat local ebsco discovery services innovative encore investigations into library web-scale discovery services | vaughan synergy serials solutions summon . overall, from a staff customization standpoint – the ability to change the interface, embed links, define facet categories, define labels, place the searchbox in a different webpage, etc., please rank each platform from best to worst. worst nd worst middle nd best best ex libris primo central oclc worldcat local ebsco discovery services innovative encore synergy serials solutions summon . regardless of a best to worst ranking, please indicate if the products were, overall, acceptable or unacceptable to you from the staff customization standpoint. unacceptable acceptable ex libris primo central oclc worldcat local ebsco discovery services innovative encore synergy serials solutions summon part iv. summary questions . overall, from a content coverage, user functionality, and staff customization standpoint, please rank each product from best to worst. worst nd worst middle nd best best ex libris primo central oclc worldcat local ebsco discovery services innovative encore synergy serials solutions summon information technology and libraries | march . regardless of a best to worst ranking, please indicate if the products were, overall, acceptable or unacceptable to you from the overall standpoint of content coverage, user functionality, and staff customization standpoint. unacceptable acceptable ex libris primo central oclc worldcat local ebsco discovery services innovative encore synergy serials solutions summon part v. additional thoughts . please share any additional thoughts you have on ex libris primo central. (freetext box) . please share any additional thoughts you have on oclc worldcat local. (freetext box) . please share any additional thoughts you have on ebsco discovery services. (freetext box) . please share any additional thoughts you have on innovative encore synergy. (freetext box) . please share any additional thoughts you have on serials solutions summon. (freetext box) investigations into library web-scale discovery services | vaughan appendix e. discovery task force: early adopter reference questions author’s note: appendix e originally appeared in the january library technology reports: web scale discovery services as chapter , “questions to consider.” part background . how long have you had your discovery service available to your end users? (what month and year did it become generally available to your primary user population, and linked to your public library website). . after you had selected a discovery service, approximately how long was the implementation period – how long did it take to “bring it up” for your end‐users and make it available (even if in ‘beta’ form) on your library website? . what have you named your discovery service, and is it the ‘default’ search service on your website at this point? in other words, regardless of other discovery systems (ils, digital collection management system, ir, etc.), has the new discovery service become the default or primary search box on your website? part content: article level content coverage & scope “article level content” = articles from academic journals, articles from mainstream journals, newspaper content, conference proceedings, open access content . in terms of article level content, do you feel the preindexed, preharvested central index of the discovery platform adequately covers a majority of the titles important to your library’s collection and focus? . have you observed any particular strengths in terms of subject content in any of the three major overarching areas -- humanities, social sciences, sciences? . have you observed any big, or appreciable, gaps in any of the three major overarching areas – humanities, social sciences, sciences? . have you observed that the discovery service leans toward one or a few particular content types (e.g. peer reviewed academic journal content; mainstream journal content; newspaper article content; conference proceedings content; academic open access content)? . are there particular publishers whose content is either not incorporated, (or not adequately incorporated), into the central index, that you’d like to see included (e.g. elsevier journal content)? . have you received any feedback, positive or negative, from your institution’s faculty, related to the content coverage within the discovery service? . taking all of the above questions into consideration, are you happy, satisfied, or dissatisfied with the scope of subject content, and formats covered, in the discovery platform’s central index? . in general, are you happy with the level of article level metadata associated with the returned information technology and libraries | march citation level results (that is, before one retrieves the complete full text). in other words, the product may incorporate basic citation level metadata (e.g. title, author, publication info), or it may include additional enrichment content, such as abstracts, author supplied keywords, etc. overall, how happy do you sense your library staff is with the quality and amount of metadata provided for a “majority” of the article level content indexed in the system? part content: your local library resources . it’s presumed that your local library ils bib records have been harvested into the discovery solution. do you have any other local “homegrown” collections – hosted by other systems at your library or institution – whose content has been harvested into the discovery solution? examples would include digital collection content, institutional repository content, library subject guide content, or other specialized, homegrown local database content. if so, please briefly describe the content – focus of collection, type of content (images, articles, etc.), and a ballpark number of items. if no local collections other than ils bib record content have been harvested, please skip to question . . [for local collections other than ils bib records]. did you use existing, vendor provided ingestors to harvest the local record content (i.e. ingestors to transfer the record content, apply any transformations and normalizations to migrate the local content to the underlying discovery platform schema)? or did you develop your own ingestors from scratch, or using a toolkit or application profile template provided by the vendor? . [for local collections other than ils bib records]. did you need extensive assistance from the discovery platform vendor to help harvest any of your local collections into the discovery index? if so, regardless of whether the vendor offered this assistance for free or charged a fee, were you happy with the level of service received from the vendor? . do you feel your local content (including ils bib records) is adequately “exposed” during a majority of searches? in other words, if your local harvested content equaled a million records, and the overall size of the discovery platform index was a hundred million records, do you feel your local content is “lost” for a majority of end user searches, or adequately exposed? part interface: general satisfaction level . overall, how satisfied are you and your local library colleagues with the discovery service’s interface? . do you have any sense of how satisfied faculty at your institution are with the discovery service’s interface? have you received any positive or negative comments from faculty related to the interface? . do you have any sense of how satisfied your (non-faculty) end-users are with the discovery service’s interface? have you received any positive or negative comments from users related to the interface? . have you conducted any end-user usability testing related to the discovery service? if so, can you provide the results, or otherwise some general comments on the results of these tests? . related to searching, are you happy with the relevance of results returned by the discovery service? have you noticed any consistent “goofiness,” or surprises with the returned results? if you could make a investigations into library web-scale discovery services | vaughan change in the relevancy arena, what would it be, if anything? part interface: local customization . has your library performed what you might consider any “major customization” to the product? or has it primarily been customizations such as naming the service, defining hyperlinks and the color scheme? if you’ve done more extensive customization, could you please briefly describe, and was the product architecture flexible enough to allow you to do what you wanted to do (also see question below, which is related). . is there any particular feature or function that is missing or non-configurable within the discovery service that you wish were available? . in general, are you happy with the “openness” or “flexibility” of the system in terms of how customizable it is by your library staff? part : final thoughts . overall, do you feel your selection of this vendor’s product was a good one? do you sense that your users – students and faculty – have positively received the product? . have you conducted any statistics review or analysis (through the discovery service statistics, or link resolver statistics, etc.) that would indicate or at least suggest that the discovery service has improved the discoverability of some of your materials (whether local library materials or remotely hosted publisher content). . if you have some sense of the competition in the vendor discovery marketplace, do you feel this product offers something above and beyond the other competitors in the marketplace? if so, what attracted you to this particular product, what made it stand out? information technology and libraries | march appendix f. discovery task force: initial vendor investigation questions section : general / background questions . customer install base how many current customers do you have that have which have implemented the product at their institution? (the tool is currently available to users / researchers at that institution) how many additional customers have committed to the product? how many of these customers fall within our library type (e.g. higher ed academic, public, k- )? . references can you provide website addresses for live implementations which you feel serve as a representative model matching our library type? can you provide references – the name and contact information for the lead individuals you worked with at several representative customer sites which match our library type? . pricing model, optional products describe your pricing model for a library type such as ours, including initial upfront costs and ongoing costs related to the subscription and technical support. what optional add-on services or modules (federated search, recommender services, enrichment services) do you market which we should be aware of, related to and able to be integrated with your web scale discovery solution? . technical support and troubleshooting briefly describe options customers have, and hours of availability, for reporting mission critical problems; and for reporting observed non mission-critical glitches. briefly describe any consulting services you may provide above and beyond support services offered as part of the ongoing subscription. (e.g. consulting services related to harvesting of a unique library resource for which an ingest/transform/normalize routine does not already exist). is there a process for suggesting enhancement requests for potential future incorporation into the product? . size of the centralized index. how many periodical titles does your preharvested, centralized index encompass? how many indexed items? . statistics. please describe what you feel are some of the more significant use, management or content related statistics available out-of-the-box with your system. investigations into library web-scale discovery services | vaughan are the statistics counter compliant? . ongoing maintenance activities, local library staff. for instances where the interface and discovery service is hosted on your end, please describe any ongoing local library maintenance activities associated with maintaining the service for the local library’s clientele (e.g. maintenance of the link resolver database; ongoing maintenance associated with periodic local resource harvest updates; etc.) section : local library resources . metadata requirements and existing ingestors. what mandatory record fields for a local resource has to exist for the content to be indexed and discoverable within your platform (title, date)? please verify that your platform has existing connectors -- ingest/transform/normalize tools and transfer mechanisms and/or application profiles for the following schema used by local systems at our library (e.g. marc bibliographic records; unqualified / qualified dublin core, ead, etc.) please describe any standard tools your discovery platform may offer to assist local staff in crosswalking between the local library database schema and the underlying schema within your platform. our library uses the abc digital collection management software. do you have any existing customers who also utilize this platform, whose digital collections have been harvested and are now exposed in their instance of the discovery product? our library uses the abc institutional repository software. do you have any existing customers who also utilize this platform, whose digital collections have been harvested and are now exposed in their instance of the discovery product? . resource normalization. is content for both local and remote content normalized to a single schema? if so, please offer comments on how local and remote (publisher/aggregator) content is normalized to this single underling schema. to what degree can collections from different sources have their own unique field information which is displayed and/or figures into the relevancy ranking algorithm for retrieval purposes. . schedule. for records hosted in systems at the local library, how often do you harvest information to account for record updates, modifications, deletions? can the local library invoke a manual harvest of locally hosted resource records on a per-resource basis (e.g. from a selected resource – for example, if the library launches a new digital collection and want the records to be available in the new discovery platform shortly after they are available in our local digital collection management system, is there a mechanism to force a harvest prior to the next regularly scheduled harvest routine? after harvesting, how long does it typically take for such updates, additions, and deletions to be reflected in the searchable central index? information technology and libraries | march . policies / procedures. please describe any general policies and procedures not already addressed which the local library should be aware of as relates to the harvesting of local resources. . consortial union catalogs. can your service harvest or provide access to items within a consortial or otherwise shared catalog (e.g. the inn-reach catalog). please describe. section : publisher and aggregator indexed content . publisher/aggregator agreements: general with approximately how many publishers have you forged content agreements with? are these agreements indefinite or do they have expiration dates? have you entered into any exclusive agreements with any publishers/aggregators (i.e. the publisher/aggregator is disallowed from forging agreements with competing discovery platform vendors, or disallowed from providing the same deep level of metadata/full text for indexing purposes). . comments on metadata provided. could you please provide some general comments on the level of data provided to you, for indexing purposes, by the “majority” of major publishers/aggregators with which you have forged agreements. please describe to what degree the following elements play a role in your discovery service: a. “basic” bibliographic information (article title/journal title/author/publication information) b. subject descriptors c. keywords (author supplied?) d. abstracts (author supplied?) e. full text . topical content strength do you feel there is a particular content area that you feel the service covers especially well or leans heavily toward (e.g. humanities, social sciences, sciences). do you feel there is a particular content type that you feel the service covers very well or leans heavily toward (scholarly journal content, mainstream journal content, newspapers, conference proceedings). what subject / content areas, if any, do you feel the service may be somewhat weak? are there current efforts to mitigate these weaknesses (e.g. future publisher agreements on the horizon)? . major publisher content agreements. are there major publisher agreements that you feel are especially significant for your service? if so, which publishers, and why (e.g. other discovery platform vendors may not have such agreements with those particular providers; the amount of content was so great that it greatly augmented the size and scope of your service; etc.) investigations into library web-scale discovery services | vaughan . content considered key by local library (by publisher). following is a list of some major publishers whose content the library licenses which is considered “key.” has your company forged agreements with these publishers to harvest their materials. if so please describe in general the scope of the agreement. how many titles are covered for each publisher? what level of metadata are they providing to you for indexing purposes (e.g. basic citation level metadata – title, author, publication date; abstracts; full text). a. ex. elsevier b. ex. sage c. ex. taylor and francis d. ex. wiley / blackwell . content considered key by local library (by title). following is a list of some major journal / newspaper titles whose content the library licenses which is considered “key.” could you please indicate if your central index includes these titles, and if so, the level of indexing (e.g. basic citation level metadata – title, author, publication date; abstracts; full text). a. ex. nature b. ex. american historical review c. ex. jama d. ex. wall street journal . google books / google scholar. do any agreements exist at this time to harvest the data associated with the google books or google scholar projects into your central index? if so, could you please describe the level of indexing (e.g. basic citation level metadata – title, author, publication date; abstracts; full text). . worldcat catalog. does your service include the oclc worldcat catalog records? if so, what level of information is included? the complete record? holdings information? . e-book vendors. does your service include items from major e-book vendors? . record information. given the fact that the same content (e.g. metadata for a unique article) can be provided by multiple sources (e.g. the original publisher of the journal itself, an open access repository, a database / aggregator, another database / aggregator, etc.), please provide some general comments on how records are built within your discovery service. for example: a. you have an agreement with a particular publisher/aggregator and they agree to provide you with rich metadata for their content, perhaps even provide you with indexing they’ve already done for their content, and may even provide you with the full text for you to be able to “deep index” their content. b. you’ve got an agreement with a particular publisher who happens to be the only publisher/provider of that content. they may provide you rich info, or they may provide you rather weak info. in any case, you choose to incorporate this into your service, as they are the only provider/publisher of the info. or, information technology and libraries | march alternately, they may not be the only publisher/provider of the info, but they are the only publisher/provider you’ve currently entered into an agreement with for that content. c. for some items appearing within your service, content for those items is provided by multiple different sources whom you’ve made agreements with. in short, there will be in some/many cases of overlap for unique items, such as a particular article title. in such cases, do you create a “merged/composite/super record” -- where your service utilizes particular metadata from each of the multiple sources, creating a “strong” single record built from these multiple resources. . deduping. related to the question immediately above, please describe your services’ approach (or not) to deduplicating items in your central index. if your service incorporates content for a same unique item from more than one content provider, does your index retrieve and display multiple instances of the same title? or do you create a merged/composite/super record, and only this single record is displayed? please describe. section : open access content . open access content sources. does your service automatically include (out of the box, no additional charge) materials from open access repositories? if so, could you please list some of the major repositories included (e.g. arxiv e-prints; hindawi publishing corporation; the directory of open access journals; hathi trust materials; etc.). . open access content sources: future plans. in addition to the current open access repositories that may be included in your service, are there other repositories whose content you are planning to incorporate in the future? . exposure to other libraries’ bibliographic / digital collection / ir content. are ils bibliographic records from other customers using your discovery platform exposed for discoverability in the searchable discovery instance of another customer? are digital collection records? institutional repository records? section : relevancy ranking . relevancy determination. please describe some of the factors which comprise the determination of relevancy within your service. what elements play a role, and how heavily are they weighted for purposes of determining relevancy? . currency. please comment on how heavily currency of an item plays in relevancy determination. does currency weigh more heavily for certain content types (e.g. newspapers)? . local library influence. does the local library have any influence or level of control over the relevancy algorithm? can they choose to “bump up” particular items for a search? please describe. . local collection visibility. could you please offer some comments on how local content (e.g. ils bibliographic records; digital collections) remains visible and discoverable within the larger pool of content indexed by your service? for example, local content may measures a million items, and your centralized index may cover half a billion items. investigations into library web-scale discovery services | vaughan . exposure of items with minimal metadata. some items likely have lesser metadata than other items. could you please offer some comments on how your system ensures discoverability for items with lesser or minimal metadata. . full text searching. does your service offer the capability for the user to search the fulltext of materials in your service (i.e. are they searching a full text keyword index?) if so, approximately what percentage of items within your service are “deep indexed?” . please describe how your system deals when no hits are retrieved for a search. does your system enable “best-match” retrieval – that is, something will always be returned or recommended? what elements play into this determination; how is the user prevented from having a completely “dead-end” search? section : authentication and rights management . open / closed nature of your discovery solution. does your system offer an unauthenticated view / access? please describe and offer some comments on what materials will not be discoverable/visible for an unauthenticated user. a. licensed full text b. records specifically or solely sourced from abstract and indexing databases c. full citation information (e.g. an unauthenticated user may see just a title; an authenticated user would see fuller citation information) d. enrichment information (such as book image covers, table of contents, abstracts, etc.) e. other . exposure of non-licensed resource metadata. if one weren’t to consider and take into account any e-journal/publisher package/database subscriptions & licenses the local library pays for, is there a base index of citation information that’s exposed and available to all subscribers of your discovery service? this may include open access materials, and/or bibliographic information for some publisher / aggregator content (which often requires a local library license to access the full text). please describe. would a user need to be authenticated to search (and retrieve results from) this “base index?” approximately how large is this “base index” which all customers may search, regardless of local library publisher/aggregator subscriptions. . rights management. please discuss how rights management is initialized and maintained in your system, for purposes of determining whether a local library user should have access to the full text (or otherwise “full resolution” if a library doesn’t license the fulltext – such as resolution to a detailed citation/abstract). information technology and libraries | march our library uses the abc link resolver. our library uses the abc a-z journal listing service. our library uses the abc electronic resource management system. is your discovery solution compatible with one/all of these systems for rights management purposes? is one approach preferable to the other, or does your approach explicitly depend on one of these particular services? section : user interface . openness to local library customization. please describe how “open” your system is to local library customization. for example, please comment on the local library’s ability to a. rename the service b. customize the header and footer hyperlinks / color scheme c. choose which facet clusters appear d. define new facet clusters e. embed the search box in other venues f. create canned, pre-customized searches for an instance of the search box g. define and promote a collection, database, or item such that it appears at the top or on the first page of any search i. develop custom “widgits” offering extra functionality or download “widgits” from an existing user community (e.g. image retrieval widgits such as flickr integration; library subject guide widgits such as libguides integration; etc. j. incorporate links to external enriched content (e.g. google book previews; amazon.com item information) k. other . web . social community features. please describe some current web . social features present in your discovery interface (e.g. user tagging, ratings, reviews, etc.). what, if any, plans do you have to offer or expand such functionality in future releases? . user accounts. does your system offer user accounts? if so, are these mandatory or optional? what services does this user account provide? a. save a list of results to return to at a later time? investigations into library web-scale discovery services | vaughan b. save canned queries for later searching? c. see a list of recently viewed items? d. perform typical ils functions such as viewing checked out items / renewals / holds? e. create customized rss feeds for a search . mobile interface. please describe the mobile interfaces available for your product. is it a browser based interface optimized for smallscreen devices? is it a dedicated iphone, android, or blackberry based executable application? . usability testing. briefly describe how your product incorporates published, established “best practices” in terms of a customer focused, usable interface. what usability testing have your performed and/or do you conduct on an ongoing basis? have any other customers that have gone live with your service completed usability testing that you’re aware of? information technology and libraries | march appendix g: vendor websites and example implementations oclc worldcat local www.oclc.org/us/en/worldcatlocal/default.htm example implementations: lincoln trails library system www.lincolntrail.info/linc.html university of delaware www.lib.udel.edu university of washington www.lib.washington.edu willamette university http://library.willamette.edu serials solutions summon www.serialssolutions.com/summon example implementations: dartmouth college www.dartmouth.edu/~library/home/find/summon drexel university www.library.drexel.edu university of calgary http://library.ucalgary.ca western michigan university http://wmich.summon.serialssolutions.com ebsco discovery services www.ebscohost.com/discovery example implementations: james madison university www.lib.jmu.edu mississippi state university http://library.msstate.edu northeastern university www.lib.neu.edu university of oklahoma http://libraries.ou.edu investigations into library web-scale discovery services | vaughan innovative interfaces encore synergy encoreforlibraries.com/tag/encore-synergy example implementations: university of nebraska-lincoln http://encore.unl.edu/iii/encore/home?lang=eng university of san diego http://sallypro.sandiego.edu/iii/encore/home?lang=eng scottsdale public library http://encore.scottsdaleaz.gov/iii/encore/home?lang=eng sacramento public library http://find.saclibrarycatalog.org/iii/encore/home?lang=eng ex libris primo central www.exlibrisgroup.com/category/primocentral example implementations: (note: example implementations are listed in alphabetical order. some implementations are more open to search by an external audience, based on configuration decisions at the local library level.) brigham young university scholarsearch www.lib.byu.edu (note: choose all-in-one search) northwestern university http://search.library.northwestern.edu vanderbilt university discoverlibrary http://discoverlibrary.vanderbilt.edu (note: choose books, media, and more) yonsei university (korea) wisearch: articles + library holdings http://library.yonsei.ac.kr/main/main.do (note: choose the articles + library holdings link. the interface is available in both korean and english; to change to english, select english at the top right of the screen after you have conducted a search and are within the primo central interface) information technology and libraries | march appendix h. vendor visit questions content . please speak to how well you feel your product stacks up against the competition in terms of the licensed full-text / citation content covered by your product. based on whatever marketplace or other competitive analysis you may have done, do you feel the agreements you’ve made with publishers equal, exceed, or trail the agreements other competitors have made? . from the perspective of an academic library serving undergraduate and graduate students as well as faculty, do you feel that there are particular licensed content areas your product covers very well (e.g. humanities, social sciences, sciences). do you feel there are areas which you need to build up? . what’s your philosophy going forward in inking future agreements with publishers to cover more licensed content? are there particular key publishers your index currently doesn’t include, but whom you are in active negotiations with? . we have several local content repositories, such as our digital collections in contentdm, our growing ir repository housed in bepress, and locally developed, web-searchable mysql databases. given the fact that most discovery platforms are quite new, do you already have existing customers harvesting their local collections, such as the above, into the discovery platform? have any particular, common problems surfaced in their attempts to get their local collections searchable and exposed in the discovery platform? . let’s say the library subscribes to an ejournal title – journal of animal studies -- that’s from a publisher with whom you don’t have an agreement for their metadata, and thus, supposedly, don’t index. if a student tried to search for an article in this journal – “giraffe behavior during the drought season,” what would happen? is this content still somehow indexed in your tool? would the discovery platform invoke our link resolver? please describe. . our focus is your next generation discovery platform, and not on your “traditional” federated search product which may be able to cover other resources not yet indexed in your next generation discovery platform. that said, please briefly describe the role of your federated search product vis a vis the next generation discovery platform. do you see your federated search product “going away” once more and more content is eventually indexed in your next generation discovery platform? end user interface & functionality . are there any particular or unique look and feel aspects of your interface that you feel elevate your product above your competitors? if so, please describe. . are there any particular or unique functionality aspects of your product that you feel elevate it above the competition (e.g. presearch or postsearch refinement categories, export options, etc.) . studies show that end users want very quick access to full text materials such as electronic journal articles and ebooks. what is your product’s philosophy in regards to this? does your platform, in your opinion, provide seamless, quick access to full text materials, with a minimum of confusion? please describe. investigations into library web-scale discovery services | vaughan related to this, does your platform de-dupe results, or is the user presented with a list of choices for a single, particular journal article they are trying to retrieve? in addition, please describe a bit how your relevancy ranking works for returned results. what makes an item appear first or on the first page of results? . please describe how “well” your product integrates with the library’s opac (in our case, innovative’s millennium opac). what information about opac holdings can be viewed directly in the discovery platform w/o clicking into the catalog and opening a new screen (e.g. call #, availability, enriched content such as table of contents or book covers?) in addition, our opac uses “scopes” which allow a user – if they choose – to limit at an outset (prior to a search being conducted) what collection they are searching. in other words, these scopes are location based, not media type based. for our institution, we have a scope for the main library, one for each of our three branch libraries, and a scope for the entire unlv collection. would your system be able to incorporate or integrate these pre-existing scopes in an advanced search mode? and/or, could these location based scopes appear as facets which a user could use to drill down a results list? . what is your platform’s philosophy in terms of “dead end searches.” does such a thing exist with your product? please describe what happens if a user a.) misspells a word b.) searches for a book or journal title / article that our library doesn’t own/license, but that we could acquire through interlibrary loan. staff “control” over the end user interface . how “open” is your platform to customization or interface design tweaks desired by the library? are there any particular aspects that the library can customize with your product that you feel elevate it above your competitors (e.g. defining facet categories; completely redesigning the end-user interface with colors, links, logos; etc.)? what are the major things customizable by the library, and why do you think this is something important that your product offers. . how “open” is your platform to porting over to other access points? in other words, provided appropriate technical skills exist, can we easily embed the search box for your product into a different webpage? could we create a “smaller,” more streamlined version of your interface for smartphone access? overarching question . in summary, what are some of the chief differentiators of your product from the competition? why is your product the best and most worthy of serious consideration? abstract introduction why web-scale discovery? q: if you could provide one piece of advice to your library, what would it be? the internal academic library perspective: genesis of the unlv libraries discovery task force the following sections of this article begin with a focus on the internal unlv library perspective—from early discussions focused on the broad topic of discovery to establishing a task force charged to identify, research, evaluate, and recommend a pot... activity: understanding web-scale activity: initial staff survey table . web-scale discovery service capabilities activity: second staff survey activity: early adopter references activity: vendor identification activity: vendor investigations activity: product development tracking activity: recommendation next steps references guest editorial clifford lynch information technology and libraries | march congratulations lita and information technology and libraries. since the early days of the internet, i’ve been continually struck by the incredible opportunities that it offers organizations concerned with the creation, organization, and dissemination of knowledge to advance their core missions in new and more effective ways. libraries and librarians were consistently early and aggressive in recognizing, seizing, and advocating for these opportunities, though they’ve faced—and continue to face—enormous obstacles ranging from copyright laws to the amazing inertia of academic traditions in scholarly communication. yet the library profession has been slow to open up access to the publications of its own professional societies, to take advantage of the greater reach and impact that such policies can offer. making these changes is not easy: there are real financial implications that suddenly seem very serious when you are a member of a board of directors, charged with a fiduciary duty to your association, and you have to push through plans to realign its finances, organizational mission, and goals in the new world of networked information. so, as a long-time lita member, i find it a great pleasure to see lita finally reach this milestone with information technology and libraries (ital) moving to fully open-access electronic distribution, and i congratulate the lita leadership for the persistence and courage to make this happen. it’s a decision that will, i believe, make the journal much more visible, and a more attractive venue for authors; it will also make it easier to use in educational settings, and to further the interactions between librarians, information scientists, computer scientists, and members of other disciplines. on a broader ala-wide level, ital now joins acrl’s college & research libraries as part of the american library association’s portfolio of open-access journals. supporting ital as an open-access journal is a very good reason indeed to be a member of lita. clifford lynch (clifford@cni.org) is executive director, coalition for networked information. mailto:clifford@cni.org president’s message: open access/open data colleen cuddy information technologies and libraries | march i am very excited to write this column. this issue of information technology and libraries (ital) marks the beginning of a new era for the journal. ital is now an open-access, electronic-only journal. there are many people to thank for this transition. the lita publications committee led by kristen antelman did a thorough analysis of publishing options and presented a thoughtful proposal to the lita board; the lita board had the foresight to push for an open-access journal even if it might mean a temporary revenue loss for the division; bob gerrity, ital editor, has enthusiastically supported this transition and did the heavy lifting to make it happen; and the lita office staff worked tirelessly for the past year to help shepherd this project. i am proud to be leading the organization during this time. to see ital go open access in my presidential year is extremely gratifying. as cliff lynch notes in his editorial, “the library profession has been slow to open up access to the publications of its own professional societies, to take advantage of the greater reach and impact that such policies can offer.” as librarians challenge publishers to pursue open-access venues, myself included, i am relieved to no longer be a hypocrite. by supporting open access we are sending a strong message to the community that we believe in the benefits of open access and we encourage other library organizations to do the same. ital will now reach a much broader and larger audience. this will benefit our authors, the organization, and the scholarship of our profession. i understand that while our members embrace open access, not everyone is pleased with an online-only journal. the number of new journals being offered electronically only is growing and i believe we are beginning to see a decline in the dual publishing model of publishers and societies offering both print and online journals. my library has been cutting back consistently on print copies of journals and this year will get only a handful of journals in print. personally, i have embraced the electronic publishing world. in fact, i held off on subscribing to the new yorker until it had an ipad subscription model! i estimate that i read percent of my books and all of my professional journals electronically. the revolution has happened for me and for many others. i know that our membership will adapt and transition their ital reading habits to our new electronic edition and i look forward to seeing this column and the entire journal in its new format. colleen cuddy (colleen.cuddy@med.cornell.edu) is lita president - and director of the samuel j. wood library and c. v. starr biomedical information center at weill cornell medical college, new york, new york. mailto:colleen.cuddy@med.cornell.edu president’s message | cuddy earlier this week saw the research works act die. librarians and researchers across the country celebrated this victory as we preserved an important open-access mandate requiring the deposition of research articles funded by the national institutes of health into pubmed central. this act threatened not just research but the availability of health information to patients and their families. as librarians, we still need to be vigilant about preserving open access and supporting open-access initiatives. i would like to draw your attention to the federal research public access act (frpaa, hr ). this act was recently introduced in the house, with a companion bill in the senate. as described by the association of research libraries, frppa would ensure free, timely, online access to the published results of research funded by eleven u.s. federal agencies. the bill gives individual agencies flexibility in choosing the location of the digital repository to house this content, as long as the repositories meet conditions for interoperability and public accessibility, and have provisions for long-term archiving. the legislation would extend and expand access to federally-funded research resources and, importantly, spur and accelerate scientific discovery. notably, this bill does not take anything away from publishers. no publisher will be forced to publish research under the bill’s provisions; any publisher can simply decline to publish the material if it feels the terms are too onerous. i encourage the library community to contact their representatives to support this bill. open access and open data are the keystones of e-science and its goals of accelerating scientific discovery. i hope that many of you will join me at the lita president’s program on june , , in anaheim. tony hey, corporate vice president of microsoft research connections and former director of the u.k.'s e-science initiative, and clifford lynch, executive director of the coalition for networked information, will discuss data-intensive scientific discovery and its implications for libraries, drawing from the seminal work the fourth paradigm. librarians are beginning to explore our role in this new paradigm of providing access to and helping to manage data in addition to bibliographic resources. it is a timely topic and one in which librarians, due to our skill set, are poised to take a leadership role. reading the fourth paradigm was a real game changer for me. it is still extremely relevant. you might consider reading a chapter or two prior to the program. it is an open-access e-book available for download from microsoft research (http://research.microsoft.com/en-us/collaboration/fourthparadigm/). i keep a copy on my ipad, right there with downloaded ital article pdfs. http://www.arl.org/pp/access/frpaa- .shtml http://research.microsoft.com/en-us/collaboration/fourthparadigm/ editor’s comments bob gerrity information technology and libraries | march welcome to the first issue of information technology and libraries (ital) as an open-access, e- only publication. as announced to lita members in early january, this change in publishing model will help ensure the long-term viability of ital by making it more accessible, more current, more relevant, and more environmentally friendly. ital will continue to feature high-quality articles that have undergone a rigorous peer-review process, but it will also begin expanding content to include more case studies, commentary, and information about topics and trends of interest to the lita community and beyond. look for a new scope statement for ital shortly. we’re pleased to include in this issue the winning paper from the lita/ex libris student writing award contest, abigail mcdermott’s overview on copyright law. we also have two lengthier-than-usual studies on library discovery services. the first, jason vaughan’s overview of his library’s investigations into web-scale discovery options, was accepted for publication more than a year ago, but due to its length did not make it into “print” until now, since we no longer face the constraints associated with the production of a print journal. the second study, by jody condit fagan and colleagues at james madison university, focuses on discovery-tool usability. jimmy ghaphery and erin white provide a timely overview of the results of their surveys on the use and management of web-based research guides. tomasz neugebauer and bin han offer a strategy and workflow for batch importing electronic theses and dissertations (etds) into an eprints repository. with the first open-access, e-only issue launched, our attention will be turned to updating and improving the ital website and expanding the back content available. our goal is to have all of the back issues of both ital and its predecessor, journal of library automation (jola), openly available from the ital site. we’ll also be exploring ways to better integrate the italica blog and the ital preprints site with the main site. suggestions and feedback are welcome, at the e-mail address below. bob gerrity (robert.gerrity@bc.edu) is associate university librarian for information technology, boston college libraries, chestnut hill, massachusetts. information technology and libraries | december within that goal are two strategies that lend them- selves to the topics including playing a role with the office for information technology policy (oitp) with regard to technology related public policy and actively participating in the creation and adoption of international standards within the library community. colby riggs (university of california–irvine) rep- resents lita on the office for information technology policy advisory committee. she also serves on the lita technology access committee, which addresses similar issues. the committee is chaired by elena m. soltau (nova southeastern university). the standards interest group is chaired by anne liebst (linda hall library of science, engineering, and technology). yan han (university of arizona) chairs the standards task force, which was charged to explore and recommend strategies and initiatives lita can implement to become more active in the creation and adoption of new technol- ogy standards that align with the library community. the task force will submit their final report before the ala midwinter meeting. for ongoing information about lita committees, interest groups, task forces, and activities being imple- mented on these and related topics, be sure to check out ala connect (http://connect.ala.org/) and the lita website (http://www.lita.org). the lita electronic dis- cussion list is there to pose questions you might have. lita members have an opportunity to advocate and participate in a leadership role as the broadband initia- tive sets the infrastructure for the next ten to fifteen years. lita members are encouraged to pursue these opportu- nities to ensure a place at the table for lita, its members, and libraries. b y now, most lita members have likely heard about the broadband technology opportunities program (btop) and the national broadband plan. the federal government is allocating grants to the states to develop their broadband infrastructure, and libraries are receiving funding to implement and expand computing in their local facilities. by september , , the national telecommunications and information administration (ntia) will have made all btop awards. information about these initiatives can be found at www .ntia.doc.gov (btop), www.broadband.gov (national broadband plan), and www.ala.org/ala/aboutala/offices/oitp/index.cfm (ala office for information technology policy). on september , , a public forum was held in silicon valley to discuss e-rate modernization and innovation in education. the conversation addressed the need to prepare schools and public libraries for broad- band. information about the forum is archived at blog .broadband.gov. established in , the e-rate program has provided funding for k– schools and public librar- ies for telecommunications and internet access. the program was successful in a dial-up world. it is time to now address broadband access which is not ubiquitous on a national basis. while the social norm suggests that technology is everywhere and everyone has the skills to use it, there is still plenty of work left to do to ensure that people can use technology and compete in an increasingly digital and global world. how does lita participate? the new strategic plan includes an advocacy and policy goal that calls for lita to advocate for and participate in the adoption of legislation, policies, technologies, and standards that promote equitable access to information and technology. karen j. starr (kstarr@nevadaculture.org) is lita president – and assistant administrator for library and develop- ment services, nevada state library and archives, carson city. karen j. starr president’s message: btop, broadband, e-rate, and lita editorial | truitt ■■ the space in between in my opinion, ital has an identity crisis. it seems to try in many ways to be scholarly like jasist, but lita simply isn’t as formal a group as asist. on the other end of the spectrum, code lib is very dynamic, infor- mal and community-driven. ital kind of flops around awkwardly in the space in between. —comment by a respondent to ital’s reader survey, december last december and january, you, the readers of information technology and libraries were invited to participate in a survey aimed at helping us to learn your likes and dis- likes about ital, and where you’d like to see this journal go in terms of several important questions. the responses provide rich food for reflection about ital, its readers, what we do well and what we don’t, and our future directions. indeed, we’re still digesting and discussing them, nearly a year after the survey. i’d like to use some of my editorial space in this issue to introduce, provide an overview, and highlight a few of the most interesting results. i strongly encourage you to access the full survey results, which i’ve posted to our weblog italica (http:// ital-ica.blogspot.com/); i further invite you to post your own thoughts there about the survey results and their meaning. we ran the survey from mid-december to mid-january. a few responses trickled in as late as mid-february. the survey invitation was sent to the , lita personal mem- bers; nonmembers and ital subscribers (most of whom are institutions) were excluded. we ultimately received responses—including two from individuals who con- fessed that they were not actually lita members—for a response rate of . percent. thus the findings reported below reflect the views of those who chose to respond to the survey. the response rate, while not optimal, is not far from the percent that i understand lita usually expects for its surveys. as you may guess, not all respondents answered all questions, which accounts for some small discrepancies in the numbers reported. who are we? in analyzing the survey responses, one of the first things one notices is the range and diversity of ital’s reader base, and by extension, of lita’s membership. the larg- est groups of subscribers identify themselves either as traditional systems librarians ( , or . percent) or web services/development librarians ( , or . percent), with a further cohort of . percent ( ) composed of those working with electronic resources or digital projects. but more than percent ( ) come from the ranks of library directors and associate directors. nearly percent ( ) identify their focus as being in the areas of reference, cataloguing, acquisitions, or collection development. see figure . the bottom line is that more than a third of our read- ers are coming from areas outside of library it. a couple of other demographic items: ■■ while nearly six in ten respondents ( , or . percent) work in academic libraries, that still leaves a sizable number ( , or . percent) who don’t. more than percent ( ) of the total respondents come from the public library sector. ■■ nearly half ( , or . percent) of our readers indi- cated that they have been with lita for five years or fewer. note that this does not necessarily indicate the age or number of years of service of the respondents, but it’s probably a rough indicator. still, i confess that this was something of a surprise to me, as i expected larger numbers of long-time members. and how do the numbers shake out for us old geezers? the – and greater-than- -years cohorts each composed about percent of those responding; interestingly, only . percent ( ) answered that they’d been lita members for between and years. assuming that these numbers are an accurate reflection of lita’s membership, i can’t help but wonder about the expla- nation for this anomaly.” see figure . how are we doing? question on the survey asked readers to respond to several statements: “it is important to me that articles in ital are peer- reviewed.” more than percent ( , or . percent) answered that they either “agreed” or “strongly agreed.” “ital is timely.” more than seven in ten respondents ( , or . percent) either “agreed” or “strongly agreed” that ital is timely. only ( . percent) disagreed. as a technology-focused journal, where time-to-publication is always a sensitive issue, i expected more dissatisfaction on this question (and no, that doesn’t mean that i don’t worry about the nine percent who believe we’re too slow out of the gate). marc truitt editorial: the space in between, or, why ital matters marc truitt (marc.truitt@ualberta.ca) is associate university librarian, bibliographic and information technology services, university of alberta libraries, edmonton, alberta, canada, and editor of ital. information technology and libraries | december would likely quit lita, with narrative explanations that clearly underscore the belief that ital—especially a paper ital—is viewed by many as an important benefit of membership. the following comments are typical: ■■ “lita membership would carry no benefits for me.” ■■ “dues should decrease, though.” [from a respon- dent who indicated he or she would retain lita “i use information from ital in my work and/ or i find it intellectually stimulating.” by a nearly identical margin to that regarding timeliness, ital readers ( , or . percent) either “agreed” or “strongly agreed” that they use ital in their work or find its contents stimulating. “ital is an important benefit of lita mem- bership.” an overwhelming majority ( , or . percent) of respondents either “agreed” or “strongly agreed” with this statement. this perception clearly emerges again in responses to the questions about whether readers would drop their lita membership if we produced an electronic-only or open-access ital (see below). where should we be going? several questions sought your input about different options for ital as we move for- ward. question , for example, asked you to rank how frequently you access ital content via several channels, with the choices being “print copy received via membership,” “print copy received by your institution/library,” “electronic copy from the ital website,” or “electronic copy accessed via an aggrega- tor service to which your institution/library subscribes (e.g., ebsco).” the choice most fre- quently accessed was the print copy received via membership, at . percent ( ). question asked about your preferences in terms of ital’s publication model. of the responses, . percent ( ) indicated a preference for continuance of the present arrangement, whereby we publish both paper and electronic versions simultaneously. four in ten respondents preferred that ital move to publication in electronic version only. of those who favored continued availability of paper, the great majority ( , or . per- cent) indicated in question that they simply preferred reading ital in paper. those who advocate moving to electronic-only do so for more mixed reasons (question ), the most popular being cost-effectiveness, timeliness, and the environmen- tal friendliness of electronic publication. a final question in this section asked that you respond to the statement “if ital were to become an electronic-only publication i would continue as a dues-paying member of lita.” while a reassuring . percent ( ) of you answered in the affirmative, . percent ( ) indicated that you figure . years of lita membership figure . professional position of lita members . % ( ) . % ( ) . % ( ) . % ( ) . % ( ) . % ( ) . % ( ) . % ( ) . % ( ) . % ( ) . % ( ) . % ( ) . % ( ) . % ( ) . % ( ) . % ( ) % % % % % systems librarian (includes responsibility for ils, servers, workstat... other (please specify) library director web services/development librarian deputy/associate/assistant director reference services librarian cataloging librarian consortium/network/vendor librarian electronic resources librarian digital projects/digitization librarian student teaching faculty computing professional (non-mls) resource sharing librarian acquisitions/collection development librarian other library staff (non-mls) . % ( ) . % ( ) . % ( ) . % ( ) % % % % % years or less – years – years more than years editorial | truitt his lipstick-on-a-pig ils. somewhere else there’s a library blogger who fends off bouts of insomnia by reading “wonky” ital papers in the wee hours of the morning. and that ain’t the half of it, as they say. in short—in terms of readers, interests, and prefer- ences—“the space in between” is a pretty big niche for ital to serve. we celebrate it. and we’ll keep trying our best to serve it well. ■■ departures as i write these lines in late-september, it’s been a sad few weeks for those of us in the ital family. in mid-august, former ital editor jim kopp passed away following a battle with cancer. last week, dan marmion—jim’s suc- cessor as editor ( – )—and a dear friend to many of us on the current ital editorial board—also left us, the victim of a malignant brain tumor. i never met jim, but lita president karen starr eulogized him in a posting to lita-l on august , . i noted dan’s retirement due to illness in this space in march. i first met dan in the spring of , when he arrived at notre dame as the new associate director for information systems and digital access (i think the position was dif- ferently titled then) and, incidentally, my new boss. dan arrived only six weeks after my own start there. things at notre dame were unsettled at the time: the libraries had only the year before successfully implemented exlibris’ aleph ils, the first north american site to do so. while exlibris moved on to implementations at mcgill and the university of iowa, we at notre dame struggled with the challenges of supporting and upgrading a system then new to the north american market. it was not always easy or smooth, but throughout, dan always maintained an unflappable and collegial manner with exlibris staff and a quiet but supportive demeanor toward those of us who worked for him. i wish i could say that i understood and appreciated this better at the time, but i can’t. i still had some growing ahead of me—i’m sure that i still do. dan was there for me again as an enthusiastic refer- ence when i moved on, first to the university of houston in and then to the university of alberta three years later. in these jobs i’d like to think i’ve come to under- stand a bit better the complex challenges faced by senior managers in large research libraries; in the process, i know i’ve come to appreciate dan’s quiet, knowledge- able, and hands-off style with department managers. it is one i’ve tried (not always successfully) to cultivate. while i was still at notre dame, dan invited me to join the editorial board of information technology and libraries, a group which over the years has come to include many “friends of dan,” including judith carter (quite possibly the world’s finest managing editor), andy boze (ital’s membership] ■■ “ital is the major benefit to me as we don’t have funds for me to attend lita meetings or training sessions.” ■■ “the paper journal is really the only membership benefit i use regularly.” ■■ “actually my answer is more, ‘i don’t know.’ i really question the value of my lita membership. ital is at least some tangible benefit i receive. quite hon- estly, i don’t know that there really are other benefits of lita membership.” question asked about whether ital should con- tinue with its current delayed open-access model (i.e., the latest two issues embargoed for non-lita members), or go completely open-access. by a three-to-two margin, readers favored moving to an open-access model for all issues. in the following question that asked whether respondents would continue or terminate lita mem- bership were ital to move to a completely open-access publication model, the results were remarkably similar to those for the question linking print availability to lita membership, with the narrative comments again suggest- ing much the same underlying reasoning. in sum, the results suggest to me more satisfaction with ital than i might have anticipated; at the same time, i’ve only scratched the surface in my comments here. the narrative answers in particular—which i have touched on in only the most cursory fashion—have many things to say about ital’s “place,” suggestions for future articles, and a host of other worthy ideas. there is as well the whole area of crosstabbing: some of the questions, when analyzed with reference to the demographic answers in the beginning of the survey, may highlight entirely new aspects of the data. who, for instance, favors continuance of a paper ital, and who prefers electronic-only? but to come back to that reader’s comment about ital and “the space in between” that i used to frame this discussion (indeed, this entire column): to me, the demographic responses—which clearly show ital has a substantial readership outside of library it—suggest that that “space in between” is precisely where ital should be. we may or may not occupy that space “awkwardly,” and there is always room for improvement, although i hope we do better than “flop around”! the results make clear that ital’s readers—who would be you!—encompass the spectrum from the tech-savvy early-career reader of code lib journal (electronic-only, of course!) to the library administrator who satisfies her need for technol- ogy information by taking her paper copy of ital along when traveling. elsewhere on that continuum, there are reference librarians and catalogers wondering what’s new in library technology, and a traditional systems librarian pondering whether there is an open-source discovery solution out there that might breathe some new life into information technology and libraries | december between membership and receiving the journal. many of them appear to infer that a portion of their lita dues, then, are ear- marked for the publication and mailing of ital. sadly, this is not the case. in years past, ital’s income from advertising paid the bills and even generated additional revenue for lita coffers. today, the shoe is on the other foot because of declining advertis- ing revenue, but ital is still expected to pay its own way, which it has failed to do in recent years. but to those who reasonably believe that some portion of their dues is dedicated to the sup- port of ital, well, t’ain’t so. bothered by this? complain to the lita board. . as a point of comparison, consider the following results from the ital reader survey. respondents were asked to rank several publishing options on a scale of to (with = most preferred option and = least preferred option): ital should be published simultaneously as a print-on- paper journal and an electronic journal (n = ): = ( . %); = ( . %); = ( . %) ital should be published in an electronic form only (n = ): = ( . %); = ( . %); = ( . %) in other words, then as now, about % of readers preferred paper and electronic to electronic-only. . karen starr, “fw: [libs-or] jim kopp: celebration of life,” online posting, aug. , , lita-l, http://lists.ala. org/sympa/arc/lita-l/ - /msg .html (accessed sept. , ). . marc truitt, “dan marmion,” information technology & libraries (mar. ): , http://www.ala.org/ala/mgrps/ divs/lita/ital/ / mar/editorial_pdf.cfm (accessed sept. , ). webmaster), and mark dehmlow. while dan left ital in , i think that he left the journal a wonderful and last- ing legacy in these extremely capable and dedicated folks. my fondest memories of dan concern our shared pas- sion for model trains. i remember visiting a train show in south bend with him a couple of times, and our last time together (at the ala midwinter meeting in denver two years ago) was capped by a snowy trek with exlibris’ carl grant, another model train enthusiast, to the mecca of model railroading, caboose hobbies. three boys off to see their toys—oh, exquisite bliss! i don’t know whether ital or its predecessor jola have ever reprinted an editorial, but while searching the archives to find something that would honor both jim and dan, i found a piece that i hope speaks eloquently of their contributions and to ital’s reason for being. dan’s edito- rial, “why is ital important?” originally published in our june issue, appears again immediately following this column. i think its message and the views expressed therein by jim and dan remain as valid today as they were in . they also may help to frame my comments concerning our reader survey in the previous section. farewell, jim and dan. you will both be sorely missed. notes and references . a number of narrative answers to the survey make it clear that ital readers who are lita members perceive a link editorial | marmion dan marmioneditorial: why is ital important? editor’s note: what follows is a reprint of dan marmion’s editorial from ital , no. ( ), http://www.ala.org/ ala/mgrps/divs/lita/ital/ editorial.cfm. after reading, we ask you to consider: why does ital matter to you? post your thoughts on italica (http://ital-ica.blogspot .com/). s ome time ago i received an e-mail from a library school student, who asked me “why is [ital] important in the library profession?” i answered the question in this way: ital is important to the library profession for at least four reasons. first, while it is no longer the only publication that addresses the use of technology in the library profession, it is the oldest (dating back to , when it was founded as the journal of library automation) and, we like to think, most distinguished. second, not only do we publish on a myriad of topics that are pertinent to technology in libraries, we publish at least three kinds of articles on those subjects: pure scholarly articles that give the results of empirical research done on topics of importance to the profes- sion, communications from practitioners in the field that present real-world experiences from which other librarians can profit, and tutorials on specific subjects that teach our readers how to do useful things that will help them in their everyday jobs. the book and software reviews that are in most issues are added bonuses. third, it is the “official” publication of lita, the only professional organization devoted to the use of information technology in the library profession. fourth, it is a scholarly, peer-reviewed journal, and as such is an important avenue for many academic librar- ians whose career advancement depends in part on their ability to publish in this type of journal. in a sen- tence, then, ital is important to the library profession because it contributes to the growth of the profession and its professionals. after sending my response, i thought it would be interesting to see what some other people with close asso- ciations to the journal would add. thus i posed the same question to the editorial board and to the person who preceded me as editor. here are some of their comments: one of the many things that was not traditionally taught in library school was a systematic approach to problem solving—for somebody who needs to acquire this skill and doesn’t have a mentor handy, ital is a wonderful resource. over and over again, ital describes how a problem was identified and defined, explains the techniques used to investigate it, and details the conclusions that might fairly be drawn from the results of the investigation. few other journals so effectively model this approach. regardless of the specific subject of the article, the opportunity to see practical problem solving techniques demonstrated is always valuable. (joan frye williams) the one thing i would add to your points, and it ties into a couple of them, is that by some definitions a “profession” is one that does have a major publica- tion. as such, it is not only the “official” publication of lita but an identity focus for those professionals in this particular area of librarianship. in fact, ideally, i would like to think that’s more of a reason why ital is important than just the fact that it’s a perk of lita membership. (jim kopp) real world experiences from which other librarians would profit—to use your own words. that is my primary reason for reading it, although i take note of tutorials as well. and the occasional book review here may catch my eye as it is likely more detailed that what might appear in lj or booklist, and [i would] be more likely to purchase it for either my office or for the gen- eral collection. (donna cranmer) ital begins as the oldest and best-established journal for refereed scholarly work in library automation and information technology, a role that by itself is impor- tant to libraries and the library profession. ital goes beyond that role to add high-quality work that does not fit in the refereed-paper mold, helping librarians to work more effectively. as the official publication of america’s largest professional association for library and information technology, ital assures a broad audience for important work—and, thanks to its cost- recovery subscription pricing, ital makes that work available to nonmembers at prices far below the norm for scholarly publishing. (walt crawford) the journal serves as an historical record/documen- tation and joins its place with many other items that together record the history of mankind. a profes- sional/scholarly journal has a presumed life that lasts indefinitely. (ken bierman) in a sentence, ital is important to the profession because “communication is the key to our success.” dan marmion was editor of ital, – . this editorial was first published in the june issue of ital. information technology and libraries | december to paper. ital provides one means of fostering this communication in a format that is easily usable and recognizable. it is not the only communications format, but it fills a particular niche. (eric lease morgan) so there you have the thoughts of the editor and a few other folks as to why this journal is important. * * * why does ital matter to you? post your thoughts on italica (http://ital-ica.blogspot.com/). ital is a formal, traditional, and standardized way of sharing ideas within a specific segment of the library community. librarianship is an institutional profession. as an institution it is an organic organiza- tion requiring communication between its members. an advantage of written communication, especially paper-based written communication, is its ability to transcend space and time. a written document can communicate an idea long after the author has died and half way around the world. yes, electronic com- munication can do the same thing, but electronic com- munication is much more fragile than ideas committed editorial board thoughts | farnel t his past spring, my alma mater, the school of library and information studies (slis) at the university of alberta, restructured the it compo- nent of its mlis program. as a result, as of september , incoming students are expected to possess certain basic it skills before beginning their program. these skills include the following: ■■ comprehension of the components and operations of a personal computer ■■ microsoft windows file management ■■ proficiency with microsoft office (or similar) prod- ucts, including word processing and presentation software ■■ use of e-mail ■■ basic web browsing and searching this new requirement got me thinking: is this com- mon practice among ala-accredited library schools? if other schools are also requiring basic it skills prior to entry, how do those required by slis compare? so i thought i’d do a little investigating to see what others in “library school land” are doing. before i continue, a word of warning: this was by no means a rigorous sci- entific investigation, but rather an informal survey of the landscape. i started my investigation with ala’s directory of institutions offering accredited master’s programs. there are fifty-seven institutions listed in the directory. i visited each institution’s website and looked for pages describing technology requirements, computer-competency require- ments, and the like. if i wasn’t able to find the desired information after fifteen or twenty minutes, i would note “nothing found” and move on to the next. in the end i found some sort of list of technology or computer-competency requirements on thirty-three (approximately percent) of the websites. it may be the case that such a list exists on other sites and i didn’t find it. i should also note that five of the lists i found focus more on software and hardware than on skills in using said software and hardware. even considering these conditions, however, i was somewhat surprised at the low numbers. is it simply assumed that today’s students already have these skills? or is it expected that they will be picked up along the way? i don’t claim to know the answers, and discovering them would require a much more detailed and thorough investigation, but they are interesting questions nonetheless. once i had found the requirements, i examined them in some detail to get a sense of the kinds of skills listed. while i won’t enumerate them all, i did find the most common ones to be similar to those required by slis— basic comfort with a personal computer and proficiency with word processing and presentation software, e-mail, file management, and the internet. a few ( ) schools also list comfort with local systems (e-mail accounts, online courseware, etc.). several ( ) schools mention familiarity with basic database design and functionality, while a few ( ) list basic web design. very few ( ) mention compe- tency with security tools (firewalls, virus checkers, etc.), and just slightly more ( ) mention familiarity with web . tools like blogs, wikis, etc. while many ( ) specifi- cally mention searching under basic internet skills, few ( ) mention proficiency with opacs or other common infor- mation tools such as full-text databases. interestingly, one school has a computer programming requirement, with mentions of specific acceptable languages, including c++, pascal, java, and perl. but this is certainly the exception rather than the rule. i was encouraged that there seems to be a certain agreement on the basics. but i was a little surprised at the relative rarity of competency with wikis and blogs and all those web . tools that are so often used and talked about in today’s libraries. is this because there is still some uncertainty as to the utility of such tools in libraries? or is it because of a belief that the members of the millennial or “digital” generation are already expert in using them? i don’t know the reasons, but it is interesting to ponder nonetheless. i was also surprised that a level of informa- tion literacy isn’t listed more often, particularly given that we’re talking about slis programs. i do know, of course, that many of these skills will be developed or enhanced as students work their way through their programs, but it also seems to me that there is so much other material to learn that the more that can be taken care of beforehand, the better. librarians work in a highly technical and techno- logical environment, and this is only going to become even more the case for future generations of librarians. certainly, basic familiarity with a variety of applications and tools and comfort with rapidly changing technologies are major assets for librarians. in fact, ala recognizes the importance of “technological knowledge and skills” as core competencies of librarianship. specifically men- tioned are the following: ■■ information, communication, assistive, and related technologies as they affect the resources, service delivery, and uses of libraries and other information agencies. ■■ the application of information, communication, assistive, and related technology and tools consistent with professional ethics and prevailing service norms and applications. ■■ the methods of assessing and evaluating the sharon farnel editorial board thoughts: system requirements sharon farnel (sharon.farnel@ualberta.ca) is metadata & cata- loguing librarian at the university of alberta in edmonton, al- berta, canada. information technology and libraries | december references . university of alberta school of library and information studies, “degree requirements: master of library & information studies,” www.slis.ualberta.ca/mlis_degree_requirements.cfm (accessed aug. , ). . american library association office for accreditation, “library & information studies directory of institutions offer- ing accredited master’s programs – ,” , http:// ala.org/ala/educationcareers/education/accreditedprograms/ directory/pdf/lis_dir_ .pdf (accessed aug. , ). . american library association, “ala’s core competences of librarianship,” january , www.ala.org/ala/education careers/careers/corecomp/corecompetences/finalcorecomp stat .pdf (accessed aug. , ). specifications, efficacy, and cost efficiency of technol- ogy-based products and services. ■■ the principles and techniques necessary to identify and analyze emerging technologies and innovations in order to recognize and implement relevant techno- logical improvements. given what we know about the importance of tech- nology to librarians and librarianship, my investigation has left me with two questions: ( ) why aren’t more library schools requiring certain it skills prior to entry into their programs? and ( ) are those who do require them asking enough of their prospective students? i hope you, our readers, might ask yourselves these questions and join us on italica for what could turn out to be a lively discussion. generating collaborative systems for digital libraries | malizia, bottoni, and levialdi from previous experience and from research in software engineering. wasted effort and poor interoperability can therefore ensue, raising the costs of dls and jeopardizing the fluidity of information assets in the future. in addition, there is a need for modeling services and data structures as highlighted in the “digital library reference model” proposed by the delos eu network of excellence (also called the “delos manifesto”); in fact, the distribution of dl services over digital networks, typically accessed through web browsers or dedicated clients, makes the whole theme of interaction between users important, for both individual usage and remote collaboration. designing and modeling such interactions call for considerations pertaining to the fields of human– computer interaction (hci) and computer-supported cooperative work (cscw). as an example, scenario- based or activity-based approaches developed in the hci area can be exploited in dl design. to meet these needs we developed cradle (cooperative-relational approach to digital library environments), a metamodel-based digital library management system (dlms) supporting collaboration in the design, development, and use of dls, exploiting patterns emerging from previous projects. the entities of the cradle metamodel allow the specification of col- lections, structures, services, and communities of users (called “societies” in cradle) and partially reflect the delos manifesto. the metamodel entities are based on existing dl taxonomies, such as those proposed by fox and marchionini, gonçalves et al., or in the delos manifesto, so as to leverage available tools and knowl- edge. designers of dls can exploit the domain-specific visual language (dvsl) available in the cradle envi- ronment—where familiar entities extracted from the referred taxonomies are represented graphically—to model data structures, interfaces and services offered to the final users. the visual model is then processed and transformed, exploiting suitable templates, toward a set of specific languages for describing interfaces and services. the results are finally transformed into platform- independent (java) code for specific dl applications. cradle supports the basic functionalities of a dl through interfaces and service templates for managing, browsing, searching, and updating. these can be further specialized to deploy advanced functionalities as defined by designers through the entities of the proposed visual the design and development of a digital library involves different stakeholders, such as: information architects, librarians, and domain experts, who need to agree on a common language to describe, discuss, and negoti- ate the services the library has to offer. to this end, high-level, language-neutral models have to be devised. metamodeling techniques favor the definition of domain- specific visual languages through which stakeholders can share their views and directly manipulate representations of the domain entities. this paper describes cradle (cooperative-relational approach to digital library environments), a metamodel-based framework and visual language for the definition of notions and services related to the development of digital libraries. a collection of tools allows the automatic generation of several services, defined with the cradle visual language, and of the graphical user interfaces providing access to them for the final user. the effectiveness of the approach is illustrated by presenting digital libraries generated with cradle, while the cradle environment has been evaluated by using the cognitive dimensions framework. d igital libraries (dls) are rapidly becoming a pre- ferred source for information and documentation. both at research and industry levels, dls are the most referenced sources, as testified by the popularity of google books, google video, ieee explore, and the acm portal. nevertheless, no general model is uni- formly accepted for such systems. only few examples of modeling languages for developing dls are available, and there is a general lack of systems for designing and developing dls. this is even more unfortunate because different stakeholders are interested in the design and development of a dl, such as information architects, to librarians, to software engineers, to experts of the spe- cific domain served by the dl. these categories may have contrasting objectives and views when deploying a dl: librarians are able to deal with faceted categories of documents, taxonomies, and document classification; software engineers usually concentrate on services and code development; information architects favor effective- ness of retrieval; and domain experts are interested in directly referring to the content of interest without going through technical jargon. designers of dls are most often library technical staff with little to no formal training in software engineering, or computer scientists with little background in the research findings of hypertext infor- mation retrieval. thus dl systems are usually built from scratch using specialized architectures that do not benefit alessio malizia (alessio.malizia@uc m.es) is associate profes- sor, universidad carlos iii, department of informatics, madrid, spain; paolo bottoni (bottoni@di.uniroma .it) is associate pro- fessor and s. levialdi (levialdi@di.uniroma .it) is professor, “sa- pienza” university of rome, department of computer science, rome, italy. alessio malizia, paolo bottoni, and s. levialdi generating collaborative systems for digital libraries: a model-driven approach information technology and libraries | december a formal foundation for digital libraries, called s, based on the concepts of streams, (data) structures, (resource) spaces, scenarios, and societies. while being evidence of a good modeling endeavor, the approach does not specify formally how to derive a system implementation from the model. the new generation of dl systems will be highly dis- tributed, providing adaptive and interoperable behaviour by adjusting their structure dynamically, in order to act in dynamic environments (e.g., interfacing with the physical world). to manage such large and complex systems, a systematic engineering approach is required, typically one that includes modeling as an essential design activity where the availability of such domain-specific concepts as first-class elements in dl models will make application specification easier. while most of the disciplines related to dls—e.g., databases, information retrieval, and hypertext and multimedia —have underlying formal models that have properly steered them, little is available to formalize dls per se. wang described the structure of a dl system as a domain-specific database together with a user interface for querying the records stored in the database. castelli et al. present an approach involving multidimensional query languages for searching information in dl systems that is based on first-order logic. these works model metadata specifications and thus are the main examples of system formalization in dl environments. cognitive models for information retrieval, as used for example by oddy et al., focus on users’ information-seeking behav- ior (i.e., formation, nature, and properties of a users’ information need) and on how information retrieval sys- tems are used in operational environments. other approaches based on models and languages for describing the entities involved in a dl are the digital library definition language, the dspace data model (with the definitions of communities and workflow mod- els), the metis workflow framework, and the fedora structoid approach. e/r approaches are frequently used for modeling database management system (dbms) applications, but as e/r diagrams only model the static structure of a dbms, they generally do not deal deeply with dynamic aspects. temporal extensions add dynamic aspects to the e/r approach, but most of them are not object-oriented. the advent of object-oriented technol- ogy calls for approaches and tools to information system design resulting in object-oriented systems. these consid- erations drove research toward modeling approaches as supported by uml. however, since the uml metamodel is not yet wide- spread in the dl community, we adopted the e/r formalism and complemented it with the specification of the dynamics made available through the user interface, as described by malizia et al. using the metamodel, we have defined a dsvl, including basic entities and language. cradle is based on the entity-relationship (e/r) formalism, which is powerful and general enough to describe dl models and is supported by many tools as a metamodeling language. moreover, we observed that users and designers involved in the dl environment, but not coming from a software engineering background, may not be familiar with advanced formalism like unified modeling language (uml), but they usually have basic notions on database management systems, where e/r is largely employed. ■■ literature review dls are complex information systems involving technolo- gies and features from different areas, such as library and information systems, information retrieval, and hci. this interdisciplinary nature is well reflected in the various definitions of dls present in the literature. as far back as , licklider envisaged collections of digital versions of scanned documents accessible via interconnected com- puters. more recently, levy and marshall described dls as sets of collections of documents, together with digital resources, accessible by users in a distributed context. to manage the amount of information stored in such systems, they proposed some sort of user-assisting software agent. other definitions include not only printed documents, but multimedia resources in general. however differ- ent the definitions may be, they all include the presence of collections of resources, their organization in struc- tured repositories, and their availability to remote users through networks (as discussed by morgan). recent efforts toward standardization have been taken by public and private organizations. for example, a delphi study identified four main ingredients: an organized collection of resources, mechanisms for browsing and searching, a distributed networked environment, and a set of objec- tified services. the president’s information technology advisory committee (pitac) panel on digital libraries sees dls as the networked collections of digital text, doc- uments, images, sounds, scientific data, and software that make up the core of today’s internet and of tomorrow’s universally accessible digital repositories of all human knowledge. when considering dls in the context of distributed dl environments, only few papers have been produced, contrasting with the huge bibliography on dls in gen- eral. the dl group at the universidad de las américas puebla in mexico introduced the concept of personal and group spaces, relevant to the cscw domain, in the dl system context. users can share information stored in their personal spaces or share agents, thus allowing other users to perform the same search on the document collec- tions in the dl. the cited text by gonçalves et al. gives generating collaborative systems for digital libraries | malizia, bottoni, and levialdi education as discussed by wattenberg or zia. in the nsdl program, a new generation of services has been developed that includes support for teaching and learn- ing; this means also considering users’ activities or scenarios and not only information access. services for implementing personal content delivery and sharing, or managing digital resources and modeling collaboration, are examples of tools introduced during this program. the virtual reference desk (vrd) is emerging as an interactive service based on dls. with vrd, users can take advantage of domain experts’ knowledge and librar- ians’ experience to locate information. for example, the u.s. library of congress ask a librarian service acts as a vrd for users who want help in searching information categories or to interact with expert librarians to search for a specific topic. the interactive and collaborative aspects of activities taking place within dls facilitate the development of user communities. social networking, work practices, and content sharing are all features that influence the technol- ogy and its use. following borgmann, lynch sees the future of dls not in broad services but in supporting and facilitating “customization by community,” i.e., services tailored for domain-specific work practices. we also examined the research agenda on system- oriented issues in dls and the delos manifesto. the agenda abstracts the dl life cycle, identifying five main areas, and proposes key research problems. in particular we tackle activities such as formal modeling of dls and their communities and developing frameworks coherent with such models. at the architectural level, one point of interest is to support heterogeneous and distributed systems, in par- ticular networked dls and services. for interoperability, one of the issues is how to support and interoperate with different metadata models and standards to allow distrib- uted cataloguing and indexing, as in the open archive initiative (oai). finally, we are interested in the service level of the research agenda and more precisely in web services and workflow management as crucial features when including communities and designing dls for use over networks and for sharing content. as a result of this analysis, the cradle framework features the following: ■■ a visual language to help users and designers when visual modeling their specific dl (without knowing any technical detail apart from learning how to use a visual environment providing diagrams representa- tions of domain specific elements) ■■ an environment integrating visual modeling and code generation instead of simply providing an integrated architecture that does not hide technical details ■■ interface generation for dealing with different users relationships for modeling dl-related scenarios and activities. the need for the integration of multiple lan- guages has also been indicated as a key aspect of the dsvl approach. in fact, complex domains like dls typi- cally consist of multiple subdomains, each of which may require its own particular language. in the current implementation, the definition of dsvls exploits the metamodeling facilities of atom , based on graph-grammars. atom has been typically used for simulation and model transformation, but we adopt it here as a tool for system generation. ■■ requirements for modeling digital libraries we follow the delos manifesto by considering a dl as an organization (possibly virtual and distributed) for managing collections of digital documents (digital con- tents in general) and preserving their images on storage. a dl offers contextual services to communities of users, a certain quality of service, and the ability to apply specific policies. in cradle we leave the definition of quality of service to the service-oriented architecture standards we employ and partially model the applicable policy, but we focus here on crucial interactivity aspects needed to make dls usable by different communities of users. in particular, we model interactive activities and services based on librarians’ experiences in face-to-face communication with users, or designing exchange and integration procedures for communicating between insti- tutions and managing shared resources. while librarians are usually interested in modeling metadata across dls, software engineers aim at provid- ing multiple tools for implementing services, such as indexing, querying, semantics, etc. therefore we pro- vide a visual model useful for librarians and information architects to mimic the design phases they usually per- form. moreover, by supporting component services, we help software engineers to specify and add services on demand to dl environments. to this end, we use a service component model. by sharing a common language, users from different categories can communicate to design a dl system while concentrating on their own tasks (services development and design for software engineers and dl design for librarians and information architects). users are modeled according to the delos manifesto as dl end-users (subdivided into content creators, content consumers, and librarians), dl designers (librarians and information archi- tects), dl system administrators (typically librarians), and dl application developers (software engineers). several activities have been started on modeling domain specific dls. as an example, the u.s. national science digital library (nsdl) program promotes edu- cational dls and services for basic and advanced science information technology and libraries | december ■■ how that information is structured and organized (structural model) ■■ the behavior of the dl (service model) and the differ- ent societies of actors ■■ groups of services acting together to carry out the dl behavior (societal model) figure depicts the design approach supported by cradle architecture, namely, modeling the society of actors and services interacting in the domain-specific scenarios and describing the documents and metadata structure included with the library by defining a visual model for all these entities. the dl is built using a col- lection of stock parts and configurable components that provide the infrastructure for the new dl. this infrastruc- ture includes the classes of objects and relationships that make up the dl, and processing tools to create and load the actual library collection from raw documents, as well as services for searching, browsing, and collection main- tenance. finally, the code generation module generates tailored dl services code stubs by composing and special- izing components from the component pool. initially, a dl designer is responsible for formalizing (starting from an analysis of the dl requirements and characteristics) a conceptual description of the dl using metamodel concepts. model specifications are then fed into a dl generator (written in python for atom ), to produce a dl tailored suitable for specific platforms and requirements. after these design phases, cradle gener- ates the code for the user interface and the parts of code corresponding to services and actors interacting in the described society. a set of templates for code generation and designers ■■ flexible metadata definitions ■■ a set of interactive integrated tools for user activities with the generated dl system to sum up, cradle is a dlms aimed at supporting all the users involved in the development of a dl system and providing interfaces, data modeling, and services for user-driven generation of spe- cific dls. although cradle does not yet satisfy all requirements for a generic dl system, it addresses issues focused on developing interactive dl systems, stressing interfaces and communication between users. nevertheless, we employed standards when possible to leave it open for further specification or enhancements from the dl user community. extensive use of xml-based languages allows us to change document information depending on implemented recognition algorithms so that expert users can easily model their dl by selecting the best recognition and indexing algorithms. cradle evolves from the jdan (java-based environ- ment for document applications on networks) platform, which managed both document images and forms on the basis of a component architecture. jdan was based on xml technologies, and its modularity allowed its integra- tion in service-based and grid-based scenarios. it supported template code generation and modeling, but it required the designer to write xml specifications and edit xml schema files in order to model the dl document types and services, thus requiring technical knowledge that should be avoided to let users concentrate on their specific domains. ■■ modeling digital library systems the cradle framework shows a unique combination of features: it is based on a formal model, exploits a set of domain-specific languages, and provides automatic code generation. moreover, fundamental roles are played by the concepts of society and collaboration. cradle generates code from tools built after modeling a dl (according to the rules defined by the proposed metamodel) and performs automatic transformation and mapping from model to code to generate software tools for a given dl model. the specification of a dl in cradle encompasses four complementary dimensions: ■■ multimedia information supported by the dl (collection model) figure . cradle architecture generating collaborative systems for digital libraries | malizia, bottoni, and levialdi socioeconomic, and environment dimen- sions. we now show in detail the entities and relations in the derived metamodel, shown in figure . actor entities actors are the users of dls. actors interact with the dl through services (interfaces) that are (or can be) affected by the actors preferences and messages (raised events). in the cradle metamodel, an actor is an entity with a behavior that may concurrently generate events. communications with other actors may occur synchronously or asynchronously. actors can relate through services to shape a digital community, i.e., the basis of a dl society. in fact, communities of students, readers, or librarians interact with and through dls, generally follow- ing predefined scenarios. as an example, societies can behave as query generator services (from the point of view of the library) and as teaching, learning, and working services (from the point of view of other humans and organiza- tions). communication between actors within the same or different societies occur through message exchange. to operate, societies need shared data structures and message protocols, enacted by sending structured sequences of queries and retrieving collections of results. the actor entity includes three attributes: . role identifies which role is played by the actor within the dl society. examples of specific human roles include authors, publishers, editors, maintain- ers, developers, and the library staff. examples of nonhuman actors include computers, printers, tele- communication devices, software agents, and digital resources in general. . status is an enumeration of possible statuses for the actor: i. none (default value) ii. active (present in the model and actively generat- ing events) iii. inactive (present in the model but not generating events) iv. sleeping (present in the model and awaiting for a response to a raised event) . events describes a list of events that can be raised by the actor or received as a response message from a service. examples of events are borrow, reserve, return, etc. events triggered from digital resources include store, trash, and transfer. examples of response events are found, not found, updated, etc. have been built for typical services of a dl environment. to improve acceptability and interoperability, cradle adopts standard specification sublanguages for representing dl concepts. most of the cradle model primitives are defined as xml elements, possibly enclos- ing other sublanguages to help define dl concepts. in more detail, mime types constitute the basis for encod- ing elements of a collection. the xml user interface language (xul) is used to represent appearance and visual interfaces, and xdoclet is used in the libgen code generation module, as shown in figure . ■■ the cradle metamodel in the cradle formalism, the specification of a dl includes a collection model describing the maintained multimedia documents, a structural model of informa- tion organization, a service model for the dl behavior, and a societal model describing the societies of actors and groups of services acting together to carry out the dl behavior. a society is an instance of the cradle model defined according to a specific collaboration framework in the dl domain. a society is the highest-level component of a dl and exists to serve the information needs of its actors and to describe its context of usage. hence a dl collects, preserves, and shares information artefacts for society members. the basic entities in cradle are derived from the categorization along the actors, activities, components, figure . the cradle metamodel with the e/r formalism information technology and libraries | december a text document, including scientific articles and books, becomes a sequence of strings. the struct entity a struct is a structural element specifying a part of a whole. in dls, structures represent hypertexts, taxono- mies, relationships between elements, or containment. for example, books can be structured logically into chap- ters, sections, subsections, and paragraphs, or physically into cover, pages, line groups (paragraphs), and lines. structures are represented as graphs, and the struct entity (a vertex) contains four attributes: . document is a pointer to the document entity the structure refers to. . id is a unique identifier for a structure element. . type takes three possible values: i. metadata denotes a content descriptor, for instance title, author, etc. ii. layout denotes the associated layout, e.g., left frame, columns, etc. iii. item indicates a generic structure element used for extending the model. . values is a list of values describing the element con- tent, e.g., title, author, etc. actors interact with services in an event-driven way. services are connected via messages (send and reply) and can be sequential, concurrent, or task-related (when a ser- vice acts as a subtask of a macroservice). services perform operations (e.g., get, add, and del) on collections, producing collections of documents as results. struct elements are connected to each other as nodes of a graph representing metadata structures associated with documents. the metamodel has been translated to a dsvl, asso- ciating symbols and icons with entities and relations (see “cradle language and tools” below). with respect to the six core concepts of the delos manifesto (content, user, functionality, quality, policy, and architecture), con- tent can be modeled in cradle as collections and structs, user as actor, and functionality as service. the quality con- cept is not directly modeled in cradle, but for quality of service we support standard service architecture. policies can be partially modeled by services managing interaction between actors and collections, making it possible to apply standard access policies. from the architectural point of view, we follow the reference architecture of figure . ■■ cradle language and tools in this section we describe the selection of languages and tools of the cradle platform. to improve interoperability service entities services describe scenarios, activities, operations, and tasks that ultimately specify the functionalities of a dl, such as collecting, creating, disseminating, evaluating, organizing, personalizing, preserving, requesting, and selecting documents and providing services to humans concerned with fact-finding, learning, gathering, and exploring the content of a dl. all these activities can be described and implemented using scenarios and appear in the dl setting as a result of actors using services (thus societies). furthermore, these activities realize and shape relationships within and between societies, services, and structures. in the cradle metamodel, the service entity models what the system is required to do, in terms of actions and processes, to achieve a task. a detailed task analysis helps understand the current system and the information flow within it in order to design and allocate tasks appropriately. the service entity has four attributes: . name is a string representing a textual description of the service. . sync states whether communication is synchronous or asynchronous, modeled by values wait and nowait, respectively. . events is a list of messages that can trigger actions among services (tasks); for example, valid or notvalid in case of a parsing service. . responses contain a list of response messages that can reply to raised events; they are used as a communica- tion mechanism by actors and services. the collection entity collections are sets of documents of arbitrary type (e.g., bits, characters, images, etc.) used to model static or dynamic content. in the static interpretation, a collection defines information content interpreted as a set of basic elements, often of the same type, such as plain text. examples of dynamic content include video delivered to a viewer, ani- mated presentations, and so on. the attributes of collection are name and documents. name is a string, while documents is a list of pairs (documentname, documentlabel), the latter being a pointer to the document entity. the document entity documents are the basic elements in a dl and are modeled with attributes label and structure. label defines a textual string used by a collection entity to refer to the document. we can consider it as a document identifier, specifying a class or a type of document. structure defines the semantics and area of appli- cation of the document. for example, any textual representation can be seen as a string of characters, so that generating collaborative systems for digital libraries | malizia, bottoni, and levialdi graphs. model manipulation can then be expressed via graph grammars also specified in atom . the general process of automatic creation of coop- erative dl environments for an application is shown in figure . initially, a designer formalizes a conceptual description of the dl using the cradle metamodel concepts. this phase is usually preceded by an analysis of requirements and interaction scenarios, as seen previ- ously. model specifications are then provided to a dl code generator (written in python within atom ) to pro- duce dls tailored to specific platforms and requirements. these are built on a collection of templates of services and configurable components providing infrastructure for the new dl. the sketched infrastructure includes classes for objects (tasks), relationships making up the dl, and processing tools to upload the actual library collection from raw documents, as well as services for searching and browsing and for document collections maintenance. the cradle generator automatically generates different kinds of output for the cradle model of the cooperative dl environment, such as service and collection managers. collection managers define the logical schemata of the dl, which in cradle correspond to a set of mime types, xul and xdoclet specifications, representing digital objects, their component parts, and linking infor- mation. collection managers also store instances of their and collaboration, cradle makes extensive use of existing standard spec- ification languages. most cradle outputs are defined with xml-based formats, able to enclose other specific languages. the basic languages and corresponding tools used in cradle are the following: ■■ mime type. multipurpose internet mail extensions (mime) constitute the basis for encoding documents in cradle, supporting several file formats and types of charac- ter encoding. mime was chosen because of wide availability of mime types, and standardisation of the approach. this makes it a natural choice for dls where dif- ferent types of documents need to be managed (pdf, html, doc, etc.). moreover, mime standards for character encoding descrip- tions help keeping the cradle framework open and compliant with standards. ■■ xul. the xml user interface language (xul) is an xml-based markup language used to represent appearance and visual interfaces. xul is not a public standard yet, but it uses many existing standards and technologies, including dtd and rdf, which makes it easily readable for peo- ple with a background in web programming and design. the main benefit of xul is that it provides a simple definition of common user interface elements (widgets). this drastically reduces the software devel- opment effort required for visual interfaces. ■■ xdoclet. xdoclet is used for generating services from tagged-code fragments. it is an open-source code generation library which enables attribute-ori- ented programming for java via insertion of special tags. it includes a library of predefined tags, which simplify coding for various technologies, e.g., web services. the motivation for using xdoclet in the cradle framework is related to its approach for template code generation. designers can describe templates for each service (browse, query, and index) and the xdoclet generated code can be automatically transformed into the java code for managing the specified service. ■■ atom . atom is a metamodeling system to model graphical formalisms. starting from a metaspecifi- cation (in e/r), atom generates a tool to process models described in the chosen formalism. models are internally represented using abstract syntax figure . cooperative dl generation process with cradle framework information technology and libraries | december and ( ) the metadata operations box. the right column manages visualization and mul- timedia information obtained from documents. the basic features provided with the ui templates are docu- ment loading, visualization, metadata organization, and management. the layout template, in the collection box, manages the visualization of the documents contained in a collection, while the visualization template works according to the data (mime) type specified by the document. actually, by selecting a document included in the collection, the corresponding data file is automatically uploaded and visualized in the ui. the metadata visualization in the code template reflects the metadata structure (a tree) represented by a struct, specifying the relationship between parent and child nodes. thus the xul template includes an area (the meta- data box) for managing tree structures as described in the visual model of the dl. although the tree-like visualiza- tion has potential drawbacks if there are many metadata items, there should be no real concern with medium loads. the ui template also includes a box to perform opera- tions on metadata, such as insert, delete, and edit. users can select a value in the metadata box and manipulate the presented values. figure shows an example of a ui generated from a basic template. service templates to achieve automated code generation, we use xdoclet to specify parameters and service code generation according to such parameters. cradle can automatically annotate java files with name–value pairs, and xdoclet provides a syntax for parameter specification. code generation is classes and function as search engines for the system. services classes also are generated and are represented as attribute-oriented classes involving parts and features of entities. ■■ cradle platform the cradle platform is based on a model-driven approach for the design and automatic generation of code for dls. in particular, the dsvl for cradle has four diagram types (collection, structure, service, and actor) to describe the different aspects of a dl. in this section we describe the user interface (ui) and service templates used for generating the dl tools. in particular, the ui layout is mainly generated from the structured information provided by the document, struct, and collection entities. the ui events are managed by invoking the appropriate services according to the imported xul templates. at the service and communica- tion levels, the xdoclet code is generated by the service and actor entities, exploiting their relationships. we also show how code generation works and the advanced platform features, such as automatic service discovery. at the end of the section a running example is shown, rep- resenting all the phases involved in using the cradle framework for generating the dl tools for a typical library scenario. user interface templates the generation of the ui is driven by the visual model designed by the cradle user. specifically, the model entities involved in this process are document, struct and collection (see figure ) for the basic components and lay- out of the interfaces, while linked services are described in the appropriate templates. the code generation process takes place through transformations implemented as actions in the atom metamodel specification, where graph-grammar rules may have a condition that must be satisfied for the rule to be applied (preconditions), as well as actions to be performed when the rule is executed (postconditions). a transformation is described during the visual modeling phase in terms of conditions and corresponding actions (inserting xul language statements for the interface in the appropriate code template placeholders). the gener- ated user interface is built on a set of xul template files that are automatically specialized on the basis of the attributes and relationships designed in the visual mod- eling phase. the layout template for the user interface is divided into two columns (see figure ). the left column is made of three boxes: ( ) the collection box ( ) the metadata box, figure . an example of an automatically generated user inter- face. (a) document area; (b) collection box; (c) metadata box; (d) metadata operations box. generating collaborative systems for digital libraries | malizia, bottoni, and levialdi "msg arguments.argname"> { "" , "" "" } , }; the first two lines declare a class with a name class nameimpl that extends the class name. the xdoclet template tag xdtclass:classname denotes the name of the class in the annotated java file. all standard xdoclet template tags have a namespace starting with “xdt.” the rest of the template uses xdtfield : forallfield to iterate through the fields. for each field with a tag named msg arguments.argname (checked using xdtfield : ifhasfieldtag), it creates a subarray of strings using the values obtained from the field tag parameters. xdtfield : fieldname gives the name of the field, while xdtfield : fieldtagvalue retrieves the value of a given field tag parameter. characters that are not part of some xdoclet template tags are directly copied into the generated code. the following code segment was generated by xdoclet using the annotated fields and the above tem- plate segment: public class msgargumentsimpl extends msgarguments { public static string[ ][ ] argumentnames = new string[ ][ ]{ { "eventmsg" , " event " , " eventstring " } , { " responsemsg " , " response " , " responsestring " } , }; } similarly, we generate the getter and setter methods for each field: public get () { return ; } public void set ( string value ) { based on code templates. hence service templates are xdoclet templates for transforming xdoclet code frag- ments obtained from the modeled service entities. the basic xdoclet template manages messages between services, according to the event and response attributes described in “cradle language and tools” above. in fact, cradle generates a java application (a service) that needs to receive messages (event) and reply to them (response) as parameters for the service application. in xdoclet, these can be attached to the cor- responding field by means of annotation tags, as in the following code segments: public class msgarguments { . . . . . . /* * @msg arguments.argname name="event " desc="event_string " */ protected string eventmsg = null; /* * @msg arguments.argname name="response" * desc="response_string " */ protected string responsemsg = null; } each msg arguments.argname related to a field is called a field tag. each field tag can have multiple parameters, listed after the field tag. in the tag name msg arguments .argname, the prefix serves as the namespace of all tags for this particular xdoclet application, thus avoiding naming conflicts with other standard or customized xdoclet tags. not only fields can be annotated, but also other entities such as class and functions can have tags too. xdoclet enables powerful code generation requir- ing little or no customization (depending on how much is provided by the template). the type of code to be generated using the parameters is defined by the corre- sponding xdoclet template. we have created template files composed of java codes and special xdoclet instructions in the form of xml tags. these xdoclet instructions allow conditionals (if) and loops (for), thus providing us with expressive power close to a programming language. in the following example, we first create an array containing labels and other information for each argument: public class impl extends { public static string[ ][ ] argumentnames = new string[ ][ ] { " , value ) ; }< /xdtfield : ifhasfieldtag> this translates into the following generated code: public java.lang.string get eventmsg ( ) { return eventmsg ; } public void set eventmsg ( string value ) { setvalue ( "eventmsg" , value ) ; } public java.lang.string getresponsemsg ( ) { return getresponsemsg ; } public void setresponsemsg ( string value ) { setvalue ( " responsemsg " , value ) ; } the same template is used for managing the name and sync attributes of service entities. code generation, service discovery, and advanced features a service or interface template only describes the solu- tion to a particular design problem—it is not code. consequently, users will find it difficult to make the leap from the template description to a particular implemen- tation even though the template might include sample code. others, like software engineers, might have no trouble translating the template into code, but they still may find it a chore, especially when they have to do it repeatedly. the cradle visual design environment (based on atom ) helps alleviate these problems. from just a few pieces of information (the visual model), typi- cally application-specific names for actors and services in a dl society along with choices for the design trade- offs, the tool can create class declarations and definitions implementing the template. the ultimate goal of the modeling effort remains, however, the production of reliable and efficiently executable code. hence a code generation transformation produces interface (xul) and service (java code from xdoclet templates) code from the dl model. we have manually coded xul templates specifying the static setup of the gui, the various widgets and their layout. this must be complemented with code gener- ated from a dl model of the systems dynamics coded into services. while other approaches are possible, we employed the solution implemented within the atom environment according to its graph grammar modeling approach to code generation. cradle supports a flexible iterative process for visual design and code generation. in fact, a design change might require substantial reimplementation generating collaborative systems for digital libraries | malizia, bottoni, and levialdi selecting one, the ui activates the metadata operations box—figure (d). the selected metadata node will then be presented in the lower (metadata operations) box, labeled “set metadata values,” replacing the default “none” value as shown in figure . after the metadata item is presented, the user can edit its value and save it by clicking on the “set value” button. the associated action saves the metadata information and causes its display in the intermediate box (tree-like structure), changing the visualization according to the new values. the code generation process for the do_search and front desk services is based on xdoclet templates. in particular, a message listener template is used to generate the java code for the front desk service. in fact, the front desk service is asynchronous and manages communica- tions between actors. the actors classes are generated also by using the services templates since they have attributes, events, and messages, just like the services. the do_search service code is based on the producer and consumer templates, since it is synchronous by defini- tion in the modeled scenario. a get method retrieving a collection of documents is implemented from the getter template. the routine invoked by the transformation action for struct entities performs a breadth-first exploration of the metadata tree in the visual model and attaches the cor- responding xul code for displaying the struct node in the correct position within the graph structure of the ui. collections, while a single rectangle connected to a collection represents a document entity; the circles linked to the document entity are the struct (metadata) entities. metadata entities are linked to the node rela- tionships (organized as a tree) and linked to the document entity by a metadata linktype relationship. the search service is synchro- nous (sync attribute set to “wait”). it queries the document collec- tion (get operation) looking for the requested document (using meta- data information provided by the borrow request), and waits for the result of get (a collection of docu- ments). based on this result, the service returns a boolean message “is_available,” which is then propa- gated as a response to the librarian and eventually to the student, as shown in figure . when the library designer has built the model, the transformation process can be run, executing the code generation actions associated with the entities and services represented in the model. the code generation process is based on template code snippets generated from the atom environment graph transformation engine, following the generative rules of the metamodel. we also use pre– and postconditions on application of transformation rules to have code genera- tion depend on verification of some property. the generated ui is presented in figure . on the right side, the document area is presented according to the xul template. documents are managed according to their mime type: the pdf file of the example is loaded with the appropriate adobe acrobat reader plug-in. on the left column of the ui are three boxes, according to the xul template. the collection box—figure (b)— presents the list of documents contained in the collection specified by the documents attribute of the library collec- tion entity, and allows users to interact with documents. after selecting a document by clicking on the list, it is presented in the document area—figure (a)—where it can be managed (edit, print, save, etc.). in the metadata box—figure (c)—the tree structure of the metadata is depicted according to the categoriza- tion modeled by the designer. the xul template contains all the basic layout and action features for managing a tree structure. the generated box contains the parent and child nodes according to the attributes specified in the corresponding struct elements. the user can click on the root for compacting or exploding the tree nodes; by figure . the library model, alias the model of the library society information technology and libraries | december workflow system. the release collection maintains the image files in a permanent storage, while data is written to the target database or content management software, together with xml metadata snippets (e.g., to be stored in xml native dbms). a typical configuration would have the recognition service running on a server cluster, with many data- entry services running on different clients (web browsers directly support xul interfaces). whereas current docu- ment capture environments are proprietary and closed, the definition of an xml-based interchange format allows the suitable assembly of different component-based tech- nologies in order to define a complex framework. the realization of the jdan dl system within the cradle framework can be considered as a preliminary step in the direction of a standard multimedia document managing platform with region segmentation and clas- sification, thus aiming at automatic recognition of image database and batch acquisition of multiple multimedia documents types and formats. personal and collaborative spaces a personal space is a virtual area (within the dl society) that is modeled as being owned and maintained by a user including resources (document collections, services, etc.), or references to resources, which are relevant to a task, or set of tasks, the user needs to carry out in the dl. personal spaces may thus contain digital documents in multiple media, personal schedules, visualization tools, and user agents (shaped as services) entitled with various tasks. resources within personal spaces can be allocated ■■ designing and generating advanced collaborative dl systems in this section we show the use of cradle as an analyti- cal tool helpful in comprehending specific dl phenomena, to present the complex interplays that occur between cradle components and dl concepts in a real dl appli- cation, and to illustrate the possibility of using cradle as a tool to design and generate advanced tools for dl development. modeling document images collections with cradle, the designer can provide the visual model of the dl society involved in document management and the remaining phases are automatically carried out by cradle modules and templates. we have provided the user with basic code templates for the recognition and indexing services, the data-entry plug-in, and archive release. the designer can thus simply translate the par- ticular dl society into the corresponding visual model within the cradle visual modeling editor. as a proof of concept, figure models the jdan archi- tecture, introduced in “requirements for modeling digital libraries,” exploiting the cradle visual language. the recognition service performs the automatic document rec- ognition and stores the corresponding document images, together with the extracted metadata in the archive col- lection. it interacts with the scanner actor, representing a machine or a human operator that scans paper documents. designers can choose their own segmentation method or algorithm; what is required to be compliant with the framework is to produce an xdoclet template. it stores the document images into the archive collection, with its different regions layout information according to the xml metadata schema provided by the designer. if there is at least one region marked as “not interpreted,” the data- entry service is invoked on the “not interpreted” regions. the data-entry service allows operators to evaluate the automatic classification performed by the system and edit the segmentation for indexing. operators can also edit the recognized regions with the classification engine (included in the recognition service) and adjust their values and sizes. the output of this phase is an xml description that will be imported in the indexing service for indexing (and eventually querying). the archive collection stores all of the basic informa- tion kept in jdan, such as text labels, while the indexing service, based on a multitier architecture, exploiting jboss . , has access to them. this service is responsible for turning the data fragments in the archive collection into useful forms to be presented to the final users, e.g., a report or a query result. the final stage in the recognition process could be to release each document to a content management or figure . the ui generated by cradle transforming the library model in xul and xdoclet code generating collaborative systems for digital libraries | malizia, bottoni, and levialdi and metadata, but also can share information with the various com- mittees collaborating for certain tasks. ■■ evaluation in this section we evaluate the pre- sented approach from three different perspectives: usability of the cradle notation, its expressiveness, and usability of the generated dls. usability of cradle notation we have tested it by using the well known cognitive dimensions framework for notations and visual language design. the dimensions are usually employed to evaluate the usability of a visual language or notation, or as heuristics to drive the design of innovative visual lan- guages. the significant results are as follows. abstraction gradient an abstraction is a grouping of elements to be treated as one entity. in this sense, cradle is abstraction-tolerant. it provides entities for high-level abstractions of com- munication processes and services. these abstractions are intuitive as they are visualized as the process they represent (services with events and responses) and easy to learn as their configuration implies few simple attri- butes. although cradle does not allow users to build new abstractions, the e/r formalism is powerful enough to provide basic abstraction levels. closeness of mapping cradle elements have been assigned icons to resemble their real-world counterparts (e.g., a collection is repre- sented as a set of paper sheets). the elements that do not have a correspondence with a physical object in the real world have icons borrowed from well-known notations (e.g., structs represented as graph nodes). consistency a notation is consistent if a user knowing some of its structure can infer most of the rest. in cradle, when two elements represent the same entity but can be used either as input or as output, then their shape is equal but incorporates an incoming or an outgoing message in order to differentiate them. see, for example, the icons for services or those for graph nodes representing either a according to the user’s role. for example, a conference chair would have access to conference-specific materi- als, visualization tools and interfaces to upload papers for review by a committee. similarly, we denote a group space as a virtual area in which library users (the entire dl society) can meet to conduct collaborative activities synchronously or asynchronously. explicit group spaces are created dynamically by a designer or facilitator who becomes (or appoints) the owner of the space and defines who the participants will be. in addition to direct user-to- user communication, users should be able to access library materials and make annotations on them for every other group to see. ideally, users should be able to act (and carry dl materials with them) between personal and group spaces or among group spaces to which they belong. it may also be the case, however, that a given resource is referenced in several personal or group spaces. basic functionality required for personal spaces includes capa- bilities for viewing, launching, and monitoring library services, agents, and applications. like group spaces, personal spaces should provide users with the means to easily become aware of other users and resources that are present in a given group space at any time, as well as mechanisms to communicate with other users and make annotations on library resources. we employed this personal and group space paradigm in modeling a collaborative environment in the academic conferences domain, where a conference chair can have a personal view of the document collections (resources) figure . the cradle model for the jdan framwork information technology and libraries | december of “sapienza” university of rome (undergraduate stu- dents), shown in figure , and ( ) an application employed with a project of records management in a collabora- tion between the computer science and the computer engineering department of “sapienza” university, as shown in figure . usability of the generated tools environments for single-view languages generated with atom have been extensively used, mostly in an aca- demic setting, in different areas like software and web engineering, modeling, and simulation; urban planning; etc. however, depending on the kind of the domain, generating the results may take some time. for instance, the state reachability analysis in the dl example takes a few minutes; we are currently employing a version of atom that includes petri-nets formalism where we can test the services states reachability. in general, from application experience, we note the general agreement that automated syntactical consistency support greatly simplifies the design of complex systems. finally, some users pointed out some technical limitations of the cur- rent implementation, such as the fact that it is not possible to open several views at a time. altogether, we believe this work contributes to make more efficient and less tedious the definition and main- tenance of environments for dls. our model-based approach must be contrasted with the programming- centric approach of most case tools, where the language and the code generation tools are hard-coded so that whenever a modification has to be done (whether on the language or on the semantic domain) developers have to dive into the code. ■■ conclusions and future work dls are complex information systems that integrate findings from disciplines such as hypertext, information retrieval, multimedia, databases, and hci. dl design is often a multidisciplinary effort, including library staff and computer scientists. wasted effort and poor inter- operability can therefore ensue. examining the related bibliography, we noted that there is a lack of tools or automatic systems for designing and developing coopera- tive dl systems. moreover, there is a need for modeling interactions between dls and users, such as scenario or activity-based approaches. the cradle framework fulfills this gap by providing a model-driven approach for generating visual interaction tools for dls, supporting design and automatic generation of code for dls. in particular, we use a metamodel made of different diagram types (collection, structures, service, and struct or an actor, with different colors. diffuseness/terseness a notation is diffuse when many elements are needed to express one concept. cradle is terse and not diffuse because each entity expresses a meaning on its own. error-proneness data flow visualization reduces the chance of errors at a first level of the specification. on the other hand, some mistakes can be introduced when specifying visual entities, since it is possible to express relations between source and target models which cannot generate semanti- cally correct code. however, these mistakes should be considered “programming errors more than slips,” and may be detected through progressive evaluation. hidden dependencies a hidden dependency is a relation between two elements that is not visible. in cradle, relevant dependencies are represented as data flows via directed links. progressive evaluation each dl model can be tested as soon as it is defined, without having to wait until the whole model is finished. the visual interface for the dl can be generated with just one click, and services can be subsequently added to test their functionalities. viscosity cradle has a low viscosity because making small changes in a part of a specification does not imply lots of readjustments in the rest of it. one can change prop- erties, events or responses and these changes will have only local effect. the only local changes that could imply performing further changes by hand are deleting entities or changing names; however, this would imply minimal changes (just removing or updating references to them) and would only affect a small set of subsequent elements in the same data flow. visibility a dl specification consists of a single set of diagrams fit- ting in one window. empirically, we have observed that this model usually involves no more than fifteen entities. different, independent cradle models can be simulta- neously shown in different windows. expressiveness of cradle the paper has illustrated the expressiveness of cradle by defining different entities end relationships for differ- ent dl requisites. to this end, two different applications have been considered: ( ) a basic example elaborated with the collaboration of the information science school generating collaborative systems for digital libraries | malizia, bottoni, and levialdi retrieval (reading, mass.: addison-wesley, ). . d. lucarella and a. zanzi, “a visual retrieval environ- ment for hypermedia information systems,” acm transactions on information systems ( ): – . . b. wang, “a hybrid system approach for supporting digital libraries,” international journal on digital libraries ( ): – ,. . d. castelli, c. meghini, and p. pagano, “foundations of a multidimensional query language for digital libraries,” in proc. ecdl ’ , lncs (berlin: springer, ): – . . r. n. oddy et al., eds., proc. joint acm/bcs symposium in information storage & retrieval (oxford: butterworths, ). . k. maly, m. zubair et al., “scalable digital libraries based on ncstrl/dienst,” in proc. ecdl ’ (london: springer, ): – . . r. tansley, m. bass and m. smith, “dspace as an open archival information system: current status and future direc- tions,” proc. ecdl ’ , lncs (berlin: springer, ): – . . k. m. anderson et al., “metis: lightweight, flexible, and web-based workflow services for digital libraries,” proc. rd acm/ieee-cs jcdl ’ (los alamitos, calif.: ieee computer society, ): – . . n. dushay, “localizing experience of digital content via structural metadata,” in proc. nd acm/ieee-cs jcdl ’ (new york: acm, ): – . . m. gogolla et al., “integrating the er approach in an oo environment,” proc. er, ’ (berlin: springer, ): – . . heidi gregersen and christian s. jensen, “temporal entity-relationship models—a survey,” ieee transactions on knowledge & data engineering ( ): – . . b. berkem, “aligning it with the changes using the goal-driven development for uml and mda,” journal of object technology ( ): – . . a. malizia, e. guerra, and j. de lara, “model-driven development of digital libraries: generating the user interface,” proc. mddaui ’ , http://sunsite.informatik.rwth-aachen.de/ publications/ceur-ws/vol- / (accessed oct , ). . d. l. atkins et al., “mawl: a domain-specific language for form-based services,” ieee transactions on software engineer- ing ( ): – . . j. de lara and h. vangheluwe, “atom : a tool for multi-formalism and meta-modelling,” proc. fase ’ (berlin: springer, ): – . . j. m. morales-del-castillo et al., “a semantic model of selective dissemination of information for digital libraries,” journal of information technology & libraries ( ): – . . n. santos, f. c. a. campos, and r. m. m. braga, “dig- ital libraries and ontology,” in handbook of research on digital libraries: design, development, and impact, ed. y.-l. theng et al. (hershey, pa.: idea group, ): : . . f. wattenberg, “a national digital library for science, mathematics, engineering, and technology education,” d-lib magazine no. ( ), http://www.dlib.org/dlib/october / wattenberg/ wattenberg.html (accessed oct , ); l. l. zia, “the nsf national science, technology, engineering, and mathematics education digital library (nsdl) program: new projects and a progress report,” d-lib magazine, , no. ( ), http://www.dlib.org/dlib/november /zia/ zia.html (accessed oct , ). . u.s. library of congress, ask a librarian, http://www.loc society), which describe the different aspects of a dl. we have built a code generator able to produce xul code from the design models for the dl user interface. moreover, we use template code generation integrating predefined components for the different services (xdoclet language) according to the model specification. extensions of cradle with behavioral diagrams and the addition of analysis and simulation capabilities are under study. these will exploit the new atom capabili- ties for describing multiview dsvls, to which this work directly contributed. references . a. m. gonçalves, e. a fox, “ sl: a language for declara- tive specification and generation of digital libraries,” proc. jcdl ’ (new york: acm, ): – . . l. candela et al., “setting the foundations of digital libraries: the delos manifesto,” d-lib magazine ( ), http://www.dlib.org/dlib/march /castelli/ castelli.html (accessed oct , ). . a. malizia et al., “a cooperative-relational approach to digital libraries,” proc. ecdl , lncs (berlin: springer, ): – . . e. a. fox and g. marchionini, “toward a worldwide dig- ital library,” communications of the acm ( ): – . . m. a. gonçalves et al., “streams, structures, spaces, scenarios, societies ( s): a formal model for digital libraries,” acm transactions on information systems ( ): – . . j. c. r. licklider, libraries of the future (cambridge, mass.: mit pr., ). . d. m. levy and c. c. marshall, “going digital: a look at assumptions underlying digital libraries,” communications of the acm ( ): – . . r. reddy and i. wladawsky-berger, “digital librar- ies: universal access to human knowledge—a report to the president,” , www.itrd.gov/pubs/pitac/pitac-dl- feb .pdf (accessed mar. , ). . e. l. morgan, “mylibrary: a digital library framework and toolkit,” journal of information technology & libraries ( ): – . . t. r. kochtanek and k. k. hein, “delphi study of digital libraries,” information processing management ( ): – . . s. e. howe et al., “the president’s information technology advisory committee’s february digital library report and its impact,” in proc. jcdl ’ (new york: acm, ): – . . n. reyes-farfan and j. a. sanchez, “personal spaces in the context of oa,” proc. jcdl ’ (ieee computer society, ): – . . m. wirsing, report on the eu/nsf strategic workshop on engineering software-intensive systems, , http://www.ercim. eu/eu-nsf/sis.pdf (accessed oct , ) . s. kelly and j.-p. tolvanen, domain-specific modeling: enabling full code generation (hoboken, n.j.: wiley, ). . h. r. turtle and w. bruce croft, “evaluation of an infer- ence network-based retrieval model,” acm transactions on information systems ( ): – . . r. a. baeza-yates, b. a. ribeiro-neto, modern information information technology and libraries | december .mozilla.org/en/xul (accessed mar. , ). . xdoclet, welcome! what is xdoclet? http://xdoclet .sourceforge.net/xdoclet/index.html (accessed mar. , ). . w c, extensible markup language (xml) . (fifth edition), http://www.w .org/tr/ /rec-xml- / (accessed mar. , ); w c, resource description framework (rdf), http://www.w .org/rdf/ (accessed mar. , ). . h. wada and j. suzuki, “modeling turnpike frontend system: a model-driven development framework leveraging uml metamodeling and attribute-oriented programming,” proc. models ’ , lncs (berlin: springer, ): – . . i. horrocks, constructing the user interface with statecharts (boston: addison-wesley, ). . universal discover, description, and integration oasis standard, welcome to uddi xml.org, http://uddi.xml.org/ (accessed mar. , ). . t. r. g. green and m. petre, “usability analysis of visual programming environments: a ‘cognitive dimensions frame- work,’” journal of visual languages & computing ( ): – . . j. de lara, e. guerra, and a. malizia, “model driven development of digital libraries—validation, analysis and for- mal code generation,” proc. rd webist ’ (berlin: springer, ). .gov/rr/askalib/ (accessed on mar. , ). . c. l. borgmann, “what are digital libraries? competing visions,” information processing & management ( ): – . . c. lynch, “coding with the real world: heresies and unexplored questions about audience, economics, and con- trol of digital libraries,” in digital library use: social practice in design and evaluation, ed. a. p. bishop, n. a. van house, and b. buttenfield (cambridge, mass.: mit pr., ): – . . y. ioannidis et al., “digital library information-technol- ogy infrastructure,” international journal of digital libraries ( ): – . . e. a. fox et al., “the networked digital library of theses and dissertations: changes in the university community,” jour- nal of computing higher education ( ): – . . h. van de sompel and c. lagoze, “notes from the inter- operability front: a progress report on the open archives ini- tiative,” proc. th ecdl, , lncs (berlin: springer ): – . . f. de rosa et al., “jdan: a component architecture for digital libraries,” delos workshop: digital library architectures, (padua, italy: edizioni libreria peogetto, ): – . . defined as a set of actors (users) playing roles and inter- acting with services. . mozilla developer center, xul, https://developer generating collaborative systems for digital libraries | visser and ball marijke visser and mary alice ball the middle mile: the role of the public library in ensuring access to broadband of fundamentally altering culture and society. in some circles the changes happen in real time as new web-based applications are developed, adopted, and integrated into the user’s daily life. these users are the early adopters; the internet cognoscenti. second tier users appreciate the availability of online resources and use a mix of devices to access internet content but vary in the extent to which they try the latest application or device. the third tier users also vary in the amount they access the internet but have generally not embraced its full potential, from not seeking out readily available resources to not connecting at all. regardless of the degree to which they access the internet, all of these users require basic technology skills and a robust underlying infrastructure. since the introduction of web . , the number and type of participatory web-based applications has continued to grow. many people are eagerly taking part in creating an increasing variety of web-based content because the basic tools to do so are widely available. the amateur, creating and sharing for primarily personal reasons, has the ability to reach an audience of unprecedented size. in turn, the internet audience, or virtual audience, can select from a vast menu of formats, including multimedia and print. with print resources disappearing, it is increasingly likely for an individual to only be able to access necessary material online. web-based resources are unique in that they enable an undetermined number of people, person- ally connected or complete strangers, to interact with and manipulate the content thereby creating something new with each interaction and subsequent iteration. many of these new resources and applications require much more bandwidth than traditional print resources. with the necessary technology no longer out of reach, a cross- section of society is affecting the course the twenty-first century is taking vis à vis how information is created, who can create it, and how we share it. in turn, who can access web-based content and who decides how it can be accessed become critical questions to answer. as people become more adept at using web-based tools and eager to try new applications, the need for greater broadband will intensify. the economic downturn is having a marked effect on people’s internet use. if there was a preexisting problem with inadequate access to broadband, current circumstances exacerbate it to where it needs immediate attention. access to broadband internet today increases this paper discusses the role of the public library in ensuring access to the broadband communication that is so critical in today’s knowledge-based society. it examines the culture of information in , and then asks what it means if individuals are online or not. the paper also explores current issues surrounding telecommunications and policy, and finally seeks to understand the role of the library in this highly technological, perpetually connected world. i n the last twenty years library collections have evolved from being predominantly print-based to ones that have a significant digital component. this trend, which has a direct impact on library services, has only accelerated with the advent of web . technologies and participa- tory content creation. cutting-edge libraries with next generation catalogs encourage patrons to post reviews, contribute videos, and write on library blogs and wikis. even less adventuresome institutions offer a variety of electronic databases licensed from multiple publishers and vendors. the piece of these library portfolios that is at best ignored and at worst vilified is the infrastructure that enables internet connectivity. in , broadband telecommunication is recognized as essential to access the full range of information resources. telecommunications experts articulate their concerns about the digital divide by focusing on first- and last-mile issues of bringing fiber and cable to end users. the library, particularly the public library, represents the metaphorical middle mile provid- ing the public with access to rich information content. equally important, it provides technical knowledge, sub- ject matter expertise, and general training and support to library users. this paper discusses the role of the public library in ensuring access to the broadband communication that is so critical in today’s knowledge-based society. it examines the culture of information in , and then asks what it means if individuals are online or not. the paper also explores current issues surrounding telecommunications and policy, and finally seeks to understand the role of the library in this highly technological, perpetually connected world. ■■ the culture of information information today is dynamic. as the internet contin- ues on its fast paced, evolutionary track, what we call ‘information’ fluctuates with each emerging web-based technology. theoretically a democratic platform, the internet and its user-generated content is in the process marijke visser (mvisser@alawash.org) is information technol- ogy policy analyst and mary alice ball (maryaliceball@yahoo .com) former chair, telecommunications subcommittee, office for information technology policy, american library association, washington, dc. information technology and libraries | december the geographical location of a community will also influ- ence what kind of internet service is available because of deployment costs. these costs are typically reflected in varying prices to consumers. in addition to the physical layout of an area, current federal telecommunications policies limit the degree to which incentives can be used on the local level. encouraging competition between isps, including municipal electric utilities, incumbent local exchange carriers, and national cable companies, for example, requires coordination between local needs and state and federal policies. such coordinated efforts are inherently difficult when taking into consideration the numerous differences between locales. ultimately, though, all of these factors influence the price end users must pay for internet access. with necessary infrastructure and telecommunica- tions policies in place, there are individual behaviors that also affect broadband adoption. according to the pew study, “home broadband adoption ,” percent of dial-up users are not interested in switching to broad- band. clearly there is a segment of the population that has not yet found personal relevance to high-speed access to online resources. in part this may be because they only have experience with dial-up connections. depending on dial-up gives the user an inherently inferior experi- ence because bandwidth requirements to download a document or view a website with multimedia features automatically prevent these users from accessing the same resources as a user with a high-speed connection. a dial-up user would not necessarily be aware of this differ- ence. if this is the only experience a user has it might be enough to deter broadband adoption, especially if there are other contributing factors like lack of technical com- fort or availability of relevant content. motivation to use the internet is influenced by the extent to which individuals find content personally rel- evant. whether it is searching for a job and filling out an application, looking at pictures of grandchildren, using skype to talk to a family member deployed in iraq, researching healthcare providers, updating a personal webpage, or streaming video, people who do these things have discovered personally relevant internet content and applications. understanding the potential relevance of going online makes it more likely that someone would experiment with other applications, thus increasing both the familiarity with what is available and the comfort level with accessing it. without relevant content, there is little motivation for someone not inclined to experiment with internet technology to cross what amounts to a sig- nificant hurdle to adoption. anthony wilhelm argues in a article discussing the growing digital divide that culturally relevant content is critical in increasing the likelihood that non-users will want to access web-based resources. the scope of the issue of providing culturally relevant content is underscored in the pew study, the amount of information and variety of formats avail- able to the user. in turn more content is being distributed as users create and share original content. businesses, nonprofits, municipal agencies, and educational institu- tions appreciate that by putting their resources online they reach a broader segment of their constituency. this approach to reaching an audience works provided the constituents have their own access to the materials, both physically and intellectually. it is one thing to have an internet connection and another to have the skill set nec- essary to make productive use of it. as reported in job-seeking in u.s. public libraries in , “less than % of the top u.s. retailers accept in- store paper applications.” municipal, state, and federal agencies are increasingly putting their resources online, including unemployment benefit applications, tax forms, and court documents. in addition to online documents, the report finds social service agencies may encourage clients to make appointments and apply for state jobs online. many of the processes that are now online require an ability to navigate the complexities of the internet at the same time as navigating difficult forms and websites. the combination of the two can deter someone from retrieving necessary resources or successfully completing a critical procedure. while early adopters and policy-makers debate the issues surrounding internet access, the other strata of society, knowingly or not and to varying degrees, are enmeshed in the outcomes of these ongoing discussions because their right to information is at stake. ■■ barriers to broadband access by condensing internet access issues to focus on the availability of adequate and sustainable broadband, it is possible to pinpoint four significant barriers to access: price, availability, perceived relevance, and technical skill level. the first two barriers are determined by existing telecommunications infrastructure as well as local, state, and federal telecommunications policies. the latter barri- ers are influenced by individual behaviors. both divisions deserve attention. if local infrastructure and the internet service provider (isp) options do not support broadband access to all areas within its boundaries, the result will be that some commu- nity members can have broadband services at home while others must rely on work or public access computers. it is important to determine what kind of broadband services are available (e.g., cable, dsl, fiber, satellite) and if they are robust enough to support the activities of the commu- nity. infrastructure must already be in place or there must be economic incentive for isps to invest in improving current infrastructure or in installing new infrastructure. generating collaborative systems for digital libraries | visser and ball at all. success hinges on understanding that each com- munity is unique, on leveraging its strengths, and on ameliorating its weaknesses. local government can play a significant role in the availability of broadband access. from a municipal per- spective, emphasizing the role of broadband as a factor in economic development can help define how the munici- pality should most effectively advocate for broadband deployment and adoption. gillett offers four initiatives appropriate for stimulating broadband from a local view- point. municipal governments can ■■ become leaders in developing locally relevant internet content and using broadband in their own services; ■■ adopt policies that make it easier for isps to offer broadband; ■■ subsidize broadband users and/or isps; or ■■ become involved in providing the infrastructure or services themselves. individually or in combination these four initiatives underscore the fact that government awareness of the possibilities for community growth made possible by broadband access can lead to local government sup- port for the initiatives of other local agencies, including nonprofit, municipal, or small businesses. agencies part- nering to support community needs can provide evidence to local policy makers that broadband is essential for com- munity success. once the municipality sees the potential for social and economic development, it is more likely to support policies that stimulate broadband buildout. building strong local partnerships will set the stage for the development of a sustainable broadband initiative as the different stakeholders share perspectives that take into account a variety of necessary components. when the time comes to implement a strategy, not only will different perspectives have been included, the plan will have champions to speak for it: the government, isps, public and private agencies, and community members. it is important to know which constituents are already engaged in supporting community broadband initiatives and which should be tapped. the ultimate purpose in establishing broadband internet access in a community is to benefit the individual community members, thereby stimulating local economic development. key players need to represent agencies that recognize the individual voice. a study led by strover provides an example of the importance of engaging local community leaders and agencies in developing a successful broadband access project. the study looked at thirty-six communities that received state funding to establish community technology centers (ctc). it addressed the effective use and manage- ment of ctcs and called attention to the inadequacy of supplying the hardware without community support which found that of the percent of adult americans who are not internet users, percent report they are not interested in going online. that pew can report similar information five years after the wilhelm article identifies a barrier to equitable access that has not been adequately resolved. ■■ models for sustainable broadband availability in discussing broadband, the question of what constitutes broadband inevitably arises. gillett, lehr, and osoria, in “local government broadband initiatives,” offers a functional definition: “access is ‘broadband’ if it repre- sents a noticeable improvement over standard dial-up and, once in place, is no longer perceived as the limit- ing constraint on what can be done over the internet.” while this definition works in relationship to dial-up, it is flexible enough to apply to all situations by focusing on “a noticeable improvement” and “no longer perceived as the limiting constraint” (added emphasis). ensuring sustainable broadband access necessitates anticipating future demand. short sighted definitions, applicable at a set moment in time, limit long-term viability of alterna- tive solutions. devising a sustainable solution calls for careful scru- tiny of alternative models, because the stakes are so high in the broadband debate. there are many different play- ers involved in constructing information policies. this does not mean, however, that their perspectives are mutu- ally exclusive. in debates with multiple perspectives, it is important to involve stakeholders who are aligned with the ultimate goal: assuring access to quality broadband to anyone going online. what is successful for one community may be entirely inappropriate in another; designing a successful system requires examining and comparing a range of scenarios. existing circumstances may predetermine a particular starting point, but one first step is to evaluate best prac- tices currently in place in a variety of communities to come up with a plan that meets the unique criteria of the community in question. sustainable broadband solutions need to be developed with local constituents in mind and successful solutions will incorporate the realities of cur- rent and future local technologies and infrastructure as well as local, state, and federal information policies. presupposing that the goal is to provide the commu- nity with the best possible option(s) for quality broadband access, these are key considerations to take into account when devising the plan. in addition to the technologi- cal and infrastructure issues, within a community there will be a combination of ways people access the internet. there will be those who have home access, those who need public access, and those who do not seek access information technology and libraries | december the current emphasis on universal broadband depends on selecting the best of the alternative plans according to carefully vetted criteria in order to develop a flexible and forward-thinking course of action. can we let people remain without access to robust broadband and the necessary skill set to use it effectively? no. as more and more resources critical to basic life tasks are accessible only online, those individuals that face challenges to going online will likely be socially and economically disadvantaged when compared to their online counterparts. recognition of this poten- tial for intensifying digital divide is recognized in the federal communication commission’s (fcc) national broadband plan (nbp) released in march . the nbp states six national broadband goals, the third of which is “every american should have affordable access to robust broadband service, and the means and skills to subscribe if they so choose.” research conducted for the recom- mendations in the nbp was comprehensive in scope including voices from industry, public interest, academia, and municipal and state government. responses to more than thirty public notices issued by the fcc provide evidence of wide concern from a variety of perspectives that broadband access should become ubiquitous if the united states is to be a competitive force in the twenty- first century. access to essential information such as govern- ment, public safety, educational, and economic resources requires a broadband connection to the internet. it is incumbent on government officials, isps, and community organizations to share ideas and resources to achieve a solution for providing their communities with robust and sustainable broadband. it is not necessary to have all users up to par with the early adopters. there is not a one-size-fits-all approach to wanting to be connected, nor is there a one-size-fits-all solution to providing access. what is important is that an individual can go online via a robust, high-speed connection that meets that indi- vidual’s needs at that moment. what this means for finding solutions is ■■ there needs to be a range of solutions to meet the needs of individual communities; ■■ they need to be flexible enough to meet the evolv- ing needs of these communities as applications and online content continue to change; and ■■ they must be sustainable for the long term so that the community is prepared to meet future needs that are as yet unknown. solutions to providing broadband internet access will be most successful when they are designed starting at the local level. community needs vary according to local demographics, geography, existing infrastructure, types of service providers, and how state and federal systems in place. users need a support system that high- lights opportunities available via the internet and that provides help when they run into problems. access is more than providing the infrastructure and hardware. the potential users must also find content that is cultur- ally relevant in an environment that supports local needs and expectations. strover found the most successful ctcs were located in places that “actively attracted people for other social and entertaining reasons.” in other words, the ctcs did not operate in a vacuum devoid of social context. successful adoption of the ctcs as a resource for information was dependent on the targeted population finding culturally relevant content in a supportive envi- ronment. an additional point made in the study showed that without strong community leadership, there was not significant use of the ctc even when placed in an already established community center. this has signifi- cant implications for what constitutes access as libraries plan broadband initiatives. investments in technology and a national commit- ment to ensure universal access to these new technologies in the s provide the current policy framework. as suggested by wilhelm in , to continue to move for- ward the national agenda needs to focus on updating policies to fit new information circumstances as they arise. today’s information policy debates should empha- size a similar focus. beyond accelerating broadband deployment into underserved areas, wilhelm suggests there needs to be support for training and content devel- opment that guarantees communities will actually use and benefit from having broadband deployed in their area. technology training and support for local agencies that provide the public with internet access, as well as opportunities for the individuals themselves, is essential if policies are going to actually lead to useful broadband adoption. individual and agency internet access and adoption require investment beyond infrastructure; they depend on having both culturally relevant content and the information literacy skills necessary to benefit from it. ■■ finding the right solution though it may have taken an economic crisis to bring broadband discussions into the living room, the result is causing renewed interest in a long-standing issue. many states have formed broadband task forces or councils to address the lack of adequate broadband access at the state level and, on the national front, broadband was a key component of the american recovery and reinvestment act of . the issue changes as technologies evolve but the underlying tenet of providing people access to the information and resources they need to be produc- tive members of society is the same. what becomes of generating collaborative systems for digital libraries | visser and ball difficult to measure, these kinds of social and cultural capital are important elements in ongoing debates about uses and consequences of broadband access. an ongoing challenge for those interested in the social, economic, and policy consequences of modern information networks will be to keep up with changing notions of what it means to be connected in cyberspace.” the social contexts in which a broadband plan will be enacted influence the appropriateness of different scenarios and should help guide which ones are imple- mented. engaging a variety of stakeholders will increase the likelihood of positive outcomes as community mem- bers embrace the opportunities provided by broadband internet access. it is difficult, however, to anticipate the outcomes that may occur as users become more familiar with the resources and achieve a higher level of comfort with technology. ramirez states, the “unexpected outcomes” section of many evalua- tion reports tends to be rich with anecdotes . . . . the unexpected, the emergent, the socially constructed innovations seem to be, to a large extent, off the radar screen, and yet they often contain relevant evidence of how people embrace technology and how they inno- vate once they discover its potential. community members have the most to gain from having broadband internet access. including them will increase the community’s return on its investment as they take advantage of the available resources. ramirez sug- gests that “participatory, learning, and adaptive policy approaches” will guide the community toward develop- ing communication technology policies that lead to a vibrant future for individuals and community alike. as success stories increase, the aggregation of local commu- nities’ social and economic growth will lead to a net sum gain for the nation as a whole. ■■ the role of the library public libraries play an important role in providing internet access to their community members. according to a study, the public library is the only outlet for no-fee internet access in . percent of communities nationwide; in rural communities the number goes up to . percent. beyond having desktop or, in some cases, wireless access, public libraries offer invaluable user support in the form of technical training and locally relevant content. libraries provide a secondary commu- nity resource for other local agencies who can point their clients to the library for no-fee internet access. in today’s economy where anecdotal reports show an increase in library use, particularly internet use, the role of the public policies mesh with local ordinances. local stakeholders best understand the complex interworking of their com- munity and are aware of who should be included in the decision-making process. including a local perspective will also increase the likelihood that as community needs change, new issues will be brought to the attention of policy makers and agencies who advocate for the indi- vidual community members. community agencies that already are familiar with local needs, abilities, and expectations are logical groups to be part of developing a successful local broadband access strategy. the library exemplifies a community resource whose expertise in local issues can inform infor- mation policy discussions on local, state, and federal levels. as a natural extension of library service, libraries offer the added value support necessary for many users to successfully navigate the internet. the library is an estab- lished community hub for informational resources and provides dedicated staff, technology training opportuni- ties, and no-fee public access computers with an internet connection. libraries in many communities are creating locally relevant web-based content as well as linking to other community resources on their own websites. seeking a partnership with the local library will augment a community broadband initiative. it is difficult to appreciate the impacts of current information technologies because they change so rap- idly there is not enough time to realistically measure the effects of one before it is mixed in with a new innovation. with web-based technologies there is a lag time between what those in the front of the pack are doing online and what those in the rear are experiencing. while there is general consensus that broadband internet access is critical in promoting social and economic development in the twenty-first century as is evidenced by the national purposes outlined in the nbp, there is not necessarily agreement on benchmarks for measuring the impacts. three anticipated outcomes of providing community access to broadband are ■■ civic participation will increase; ■■ communities will realize economic growth; and ■■ individual quality of life will improve. when a strategy involves significant financial and energy investments there is a tendency to want palpable results. the success of providing broadband access in a community is challenging to capture. to achieve a level of acceptable success it is necessary to focus on local communities and aggregate anecdotal evidence of incre- mental changes in public welfare and economic gain. acceptable success is subjective at best but can be usefully defined in context of local constituencies. referring to participation in the development of a vibrant culture, horrigan notes that “while inherently information technology and libraries | december isolation. an individual must possess skills to navigate the online resources. as users gain an understanding of the potential personal growth and opportunities broad- band yields, they will be more likely to seek additional online resources. by stimulating broadband use, the library will contribute to the social and economic health of the community. if the library is to extend its role as the information hub in the community by providing no-fee access to broadband to anyone who walks through the door, the local community must be prepared to support that role. it requires a commitment to encourage build out of appro- priate technology necessary for the library to maintain a sustainable internet connection. it necessitates that local communities advocate for national information and com- munication policies that are pro-library. when public policy supports the library’s efforts, the local community benefits and society at large can progress. what if the library’s own technology needs are not met? the role of the library in its community is becoming increasingly important as more people turn to it for their internet access. without sufficient revenue, the library will have a difficult time meeting this additional demand for services. in turn, in many libraries increased demand for broadband access stretches the limit of it support for both the library staff and the patrons needing help at the computers. what will be the fallout from the library not being able to provide internet services the patrons desire and require? will there be a growing skills difference between people who adopt emerging technologies and incorporate them into their daily lives and those who maintain the technological status quo? what will the social impact be of remaining off line either completely or only marginally? can the library be the bridge between those on the edge, those in the middle, and those at the end? with a strong and well articulated vision for the future, the library can be the link that provides the com- munity with sustainable broadband. ■■ conclusion the recent national focus on universal broadband access has provided an opportunity to rectify a lapse in effective information policy. whether the goal includes facilitating meaningful access continues to be more elusive. as gov- ernment, organizations, businesses, and individuals rely more heavily on the internet for sharing and receiving information, broadband internet access will continue to increase in importance. following the status quo will not necessarily lead to more people having broadband access in the long run. the early adopters will continue to stimu- late technological innovation which, in turn, will trickle down the ranks of the different user types. currently, library as a stable internet provider cannot be overesti- mated. to maintain its vital function, however, the library must also resolve infrastructure challenges of its own. because of the increased demand for access to internet resources, public libraries are finding their current broad- band services are not able to support the demand of their patrons. the issues are two-fold: increased patron use means there are often neither sufficient workstations nor broadband speeds to meet patron demand. in , about . percent of libraries reported an insufficient number of public workstations, and about . percent reported insufficient broadband speeds. to add to these already significant issues, the report indicates libraries are having trouble supporting the necessary information technology (it) because of either staff time constraints or the lack of a dedicated it staff. public libraries are facing consider- able infrastructure management issues at a time when library use is increasing. overcoming the challenges successfully will require support on the local, state, and federal level. here is where the librarian, as someone trained to become inherently familiar with the needs of her local constituency and ethically bound to provide access to a variety of information resources, needs to insert herself into the debate. librarians need to be ahead of the crowd as the voice that assures content will be readily accessible to those who seek it. today, the elemental policy issue regarding access to information via the internet hinges on connectivity to a sustainable broadband network. to promote equitable broadband access, the librarian needs be aware of the pertinent information policies in place or under consideration, and be able to anticipate those in the future. additionally, she will need to educate local policy makers about the need for broadband in their com- munity. in some circumstances, the librarian will need to move beyond her local community and raise awareness of community access issues on the state and federal level. the librarian is already able to articulate numerous issues to a variety of stakeholders and can transfer this skill to advocate for sustainable broadband strategies that will succeed in her local community. there are many strata of internet users, from those in the forefront of early adoption to those not interested in being online at all. the early adopters drive the market which responds by making resources more and more likely to be primarily available only online. as we con- tinue this trend, the social repercussions increase from merely not being able to access entertainment and news to being unable to participate in the knowledge-based society of the twenty-first century. by folding in added value online access for the community, the library helps increase the likelihood that the community will benefit from broadband being available to the library patrons and by extension to the community as a whole. to realize the internet’s full potential, access to it cannot be provided in generating collaborative systems for digital libraries | visser and ball community, the entire community benefits regardless of where and how the individuals go online. the effects of the internet are now becoming broadly social enough that there is a general awareness that the internet is not decoration on contemporary society but a challenge to it. being connected is no longer an optional luxury; to engage in the twenty-first century it is essential. access to the internet, however, is more than simple connectivity. successful access requires: an understanding of the ben- efits to going on line, technological comfort, information literacy, ongoing support and training, and the availabil- ity of culturally relevant content. people are at various levels of internet use, from those eagerly anticipating the next iteration of web-based applications to those hesitant to open an e-mail account. this user spectrum is likely to continue. though the starting point may vary depending on the applications that become important to the user in the middle of the spectrum, there will be those out in front and those barely keeping up. the implications of the pervasiveness of the internet are only beginning to be appreciated and understood. because of their involvement at the cutting edge of internet evolution, librarians can help lead the conver- sations. libraries have always been situated in neutral territory within their communities and closely aligned with the public good. librarians understand the per- spective of their patrons and are grounded in their local communities. librarians can therefore advocate effectively for their communities on issues that may not completely be understood or even recognized as matter- ing. connectivity is an issue supremely important to the library as today access to the full range of information necessitates a broadband connection. libraries have carved out a role for themselves as a premier internet access provider in the continually evolving online culture. as noted by bertot, mcclure, and jaeger, the “role of internet access provider for the community is ingrained in the social perceptions of public libraries, and public internet access has become a central part of community perceptions about libraries and the value of the library profession.” in times of both economic crisis and technological innovation, there are many unknowns. in part because of these two juxtaposed events, the role of the public library is in flux. additionally, the network of community orga- nizations that libraries link to is becoming more and more complex. it is a time of great opportunity if the library can articulate its role and frame it in relationship to broader society. evolving internet applications require increasing amounts of bandwidth and the trend is to make these bandwidth-heavy applications more and more vital to daily life. one clear path the library community can take however, the supply of internet resources is unevenly stimulating user demand and the unequal distribution of broadband access has greater potential for significant negative social consequences. staying the course and fol- lowing a haphazard evolution of broadband adoption, may, in fact, renew valid concerns about a digital divide. without an intentional and coordinated approach to developing a broadband strategy, its success is likely to fall short of expectations. the question of how to ensure that internet content is meaningful requires instituting a plan on a very local level, including stakeholders who are familiar with the unique strengths and weaknesses of their community. strover, in her article the first mile, suggests connectivity issues should be viewed from a first mile perspective where the focus is on the person accessing the internet and her qualitative experience rather than from a last mile perspective which emphasizes isp, infra- structure, and market concerns. both perspectives are talking about the same physical section of the connection network: the piece that connects the user to the network. according to strover, distinguishing between the first mile and last mile perspectives is more than an arbitrary argument over semantics. instead, a first mile perspective represents a shift “in the values and priorities that shape telecommunications policy.” by switching to a first mile perspective, connectivity issues immediately take into account the social aspects of what it means to be online. who will bring this perspective to the table? and how will we ascertain what the best approach to supporting the individual voice should be? the first mile perspective is one the library is inti- mately familiar with as an organization that traditionally advocates for the first mile of all information policies. the library is in a key position in the connectivity debate because of its inclination to speak for the user and to be aware of the unique attributes and needs of its local community. as part of its mission, the library takes into account the distinctive needs of its user community when it designs and implements its services. a natural outgrowth of this practice is to be keenly aware of the demographics of the community at large. the library can leverage its knowledge and understanding to create an even greater positive impact on the social, educational, and economic community development made possible by broadband adoption. to extend the first mile perspective analogy, in the connectivity debate, the library will play the role of the middle mile: the support system that suc- cessfully connects the internet to the consumer. while the target populations for stimulating demand for broadband are really those in the second tier of users, by advocating for the first mile perspective, the library will be advocating for equitable information policies whose implementation has bearing on the early adopters as well. by stimulating demand for broadband within a information technology and libraries | december initiatives,” . . ibid., – . . sharon strover, gary chapman, and jody waters, “beyond community networking and ctcs: access, development, and public policy,” telecommunications policy , no. / ( ): – . . ibid., . . ibid. . wilhelm, “leveraging sunken investments in communi- cations infrastructure,” . . see, for example, the virginia broadband round table (http://www.otpba.vi.virginia.gov/broadband_roundtable .shtml), the ohio broadband council (http://www.ohiobroad bandcouncil.org/), and the california broadband task force (http://gov.ca.gov/speech/ . see www.fcc.gov/recovery/ broadband/) for information on broadband initiatives in the american recovery and reinvestment act. . federal communication commission, national broad- band plan: connecting america, http://www.broadband.gov/ (accessed apr. , ). . ibid. . horrigan, “broadband: what’s all the fuss about?” . . ricardo ramirez, “appreciating the contribution of broadband ict with rural and remote communities: stepping stones toward and alternative paradigm,” the information soci- ety ( ): . . ibid., . . denise m. davis, john carlo bertot, and charles, r. mcclure, “libraries connect communities: public library funding & technology access study – ,” , http:// www.ala.org/ala/aboutala/offices/ors/plftas/ /libraries connectcommunities.pdf (accessed jan. , ). . john carlo bertot et al., “public libraries and the internet : study results and findings,” , http://www.ii.fsu.edu/ projectfiles/plinternet/ /everything.pdf (accessed jan. , ). these numbers represent an increase from the previous year’s study which suggests that libraries while trying to meet demand are not able to keep up. . ibid. . sharon strover, “the first mile,” the information society , no. ( ): – . . ibid., . . clay shirky, “here comes everybody: the power of organizing without organizations.” berkman center for inter- net & society ( ). video presentation. available at http:// cyber.law.harvard.edu/interactive/events/ / /shirky (retrieved march , ). . john carlo bertot, charles r. mcclure, and paul t. jaeger, “the impacts of free public internet access on public library patrons and communities,” library quarterly , no. ( ): , http://www.journals.uchicago.edu.proxy.ulib.iupui.edu/ doi/pdf/ . / (accessed jan. , ). is to develop its role as the middle mile connecting the increasing breadth of internet resources to the general public. the broadband debate has moved out of the background of telecommunication policy and into the center of public attention. now is the moment that calls for an information policy advocate who can represent the end user while understanding the complexity of the other stakeholder perspectives. the library undoubtedly has its own share of stakeholders, but over time it is an institution that has maintained a neutral stance within its community, thereby achieving a unique ability to speak for all parties. those who speak for the library are able to represent the needs of the public, work with a diverse group of stakeholders, and help negotiate a sustainable strategy for providing broadband internet access. references and notes . lee rainie, “ . and the internet world,” internet librar- ian , http://www.pewinternet.org/presentations/ / -and-the-internet-world.aspx (accessed mar. , ). see also john horrigan, “a typology of information and communication technology users,” , www.pewinternet.org/~/media// files/reports/ /pip_ict_typology.pdf.pdf (accessed feb. , ). . lawrence lessig, “early creative commons his- tory, my version,” video blog post, , http://lessig.org/ blog/ / /early_creative_commons_history.html (accessed jan. , ). see the relevant passage from : through : . . john horrigan, “broadband: what’s all the fuss about?” , p. , http://www.pewinternet.org/~/media/ files/reports/ /broadband% fuss.pdf.pdf (accessed feb. , ). . “job-seeking in us public libraries,” public library fund- ing & technology access study, , http://www.ala.org/ ala/research/initiatives/plftas/issuesbriefs/brief_jobs_july.pdf (accessed mar. , ). . ibid. . ibid. . sharon e. gillett, william h. lehr, and carlos osorio, “local government broadband initiatives,” telecommunications policy ( ): . . john horrigan, “home broadband adoption ,” , http://www.pewinternet.org/~/media//files/reports/ / pip_broadband_ .pdf (accessed feb. , ). . anthony g. wilhelm, “leveraging sunken investments in communications infrastructure: a policy perspective from the united states,” the information society ( ): – . . horrigan, “home broadband adoption,” . . gillett, lehr, and osorio, “local government broadband generating collaborative systems for digital libraries | hilera et al. josé r. hilera, carmen pagés, j. javier martínez, j. antonio gutiérrez, and luis de-marcos an evolutive process to convert glossaries into ontologies dictionary, the outcome will be limited by the richness of the definition of terms included in that dictionary. it would be what is normally called a “lightweight” ontol- ogy, which could later be converted into a “heavyweight” ontology by implementing, in the form of axioms, know- ledge not contained in the dictionary. this paper describes the process of creating a lightweight ontology of the domain of software engineering, starting from the ieee standard glossary of software engineering terminology. ■■ ontologies, the semantic web, and libraries within the field of librarianship, ontologies are already being used as alternative tools to traditional controlled vocabularies. this may be observed particularly within the realm of digital libraries, although, as krause asserts, objections to their use have often been raised by the digital library community. one of the core objections is the difficulty of creating ontologies as compared to other vocabularies such as taxonomies or thesauri. nonetheless, the semantic richness of an ontology offers a wide range of possibilities concerning indexing and searching of library documents. the term ontology (used in philosophy to refer to the “theory about existence”) has been adopted by the artificial intelligence research community to define a cate- gorization of a knowledge domain in a shared and agreed form, based on concepts and relationships, which may be formally represented in a computer readable and usable format. the term has been widely employed since , when berners-lee et al. envisaged the semantic web, which aims to turn the information stored on the web into knowledge by transforming data stored in every webpage into a common scheme accepted in a specific domain. to accomplish that task, knowledge must be represented in an agreed-upon and reusable computer-readable format. to do this, machines will require access to structured collections of information and to formalisms which are based on mathematical logic that permits higher levels of automatic processing. technologies for the semantic web have been devel- oped by the world wide web consortium (w c). the most relevant technologies are rdf (resource description this paper describes a method to generate ontologies from glossaries of terms. the proposed method presupposes an evolutionary life cycle based on successive transforma- tions of the original glossary that lead to products of intermediate knowledge representation (dictionary, tax- onomy, and thesaurus). these products are characterized by an increase in semantic expressiveness in comparison to the product obtained in the previous transformation, with the ontology as the end product. although this method has been applied to produce an ontology from the “ieee standard glossary of software engineering terminology,” it could be applied to any glossary of any knowledge domain to generate an ontology that may be used to index or search for information resources and documents stored in libraries or on the semantic web. f rom the point of view of their expressiveness or semantic richness, knowledge representation tools can be classified at four levels: at the basic level (level ), to which dictionaries belong, tools include defini- tions of concepts without formal semantic primitives; at the taxonomies level (level ), tools include a vocabulary, implicit or explicit, as well as descriptions of specialized relationships between concepts; at the thesauri level (level ), tools further include lexical (synonymy, hyperonymy, etc.) and equivalence relationships; and at the reference models level (level ), tools combine the previous relation- ships with other more complex relationships between concepts to completely represent a certain knowledge domain. ontologies belong at this last level. according to the hierarchic classification above, knowledge representation tools of a particular level add semantic expressiveness to those in the lowest levels in such a way that a dictionary or glossary of terms might develop into a taxonomy or a thesaurus, and later into an ontology. there are a variety of comparative studies of these tools, as well as varying proposals for systematically generating ontologies from lower-level knowledge repre- sentation systems, especially from descriptor thesauri. this paper proposes a process for generating a termino- logical ontology from a dictionary of a specific knowledge domain. given the definition offered by neches et al. (“an ontology is an instrument that defines the basic terms and relations comprising the vocabulary of a topic area as well as the rules for combining terms and relations to define extensions to the vocabulary”) it is evident that the ontology creation process will be easier if there is a vocabulary to be extended than if it is developed from scratch. if the developed ontology is based exclusively on the josé r. hilera (jose.hilera@uah.es) is professor, carmen pagés (carmina.pages@uah.es) is assistant professor, j. javier mar- tínez (josej.martinez@uah.es) is professor, j. antonio gutiér- rez (jantonio.gutierrez@uah.es) is assistant professor, and luis de-marcos (luis.demarcos@uah.es) is professor, department of computer science, faculty of librarianship and documentation, university of alcalá, madrid, spain. information technology and libraries | december configuration management; data types; errors, faults, and failures; evaluation techniques; instruction types; language types; libraries; microprogramming; operating systems; quality attributes; software documentation; soft- ware and system testing; software architecture; software development process; software development techniques; and software tools. in the glossary, entries are arranged alphabetically. an entry may consist of a single word, such as “software,” a phrase, such as “test case,” or an acronym, such as “cm.” if a term has more than one definition, the definitions are numbered. in most cases, noun definitions are given first, followed by verb and adjective definitions as applicable. examples, notes, and illustrations have been added to clarify selected definitions. cross-references are used to show a term’s relations with other terms in the dictionary: “contrast with” refers to a term with an opposite or substantially different mean- ing; “syn” refers to a synonymous term; “see also” refers to a related term; and “see” refers to a preferred term or to a term where the desired definition can be found. figure shows an example of one of the definitions of the glossary terms. note that definitions can also include framework), which defines a common data model to specify metadata, and owl (ontology web language), which is a new markup language for publishing and sharing data using web ontologies. more recently, the w c has presented a proposal for a new rdf-based markup system that will be especially useful in the con- text of libraries. it is called skos (simple knowledge organization system), and it provides a model for expressing the basic structure and content of concept schemes, such as thesauri, classification schemes, subject heading lists, taxonomies, folksonomies, and other simi- lar types of controlled vocabularies. the emergence of the semantic web has created great interest within librarianship because of the new possibili- ties it offers in the areas of publication of bibliographical data and development of better indexes and better displays than those that we have now in ils opacs. for that rea- son, it is important to strive for semantic interoperability between the different vocabularies that may be used in libraries’ indexing and search systems, and to have com- patible vocabularies (dictionaries, taxonomies, thesauri, ontologies, etc.) based on a shared standard like rdf. there are, at the present time, several proposals for using knowledge organization systems as alternatives to controlled vocabularies. for example, folksonomies, though originating within the web context, have been proposed by different authors for use within libraries “as a powerful, flexible tool for increasing the user-friendliness and inter- activity of public library catalogs.” authors argue that the best approach would be to create interoperable controlled vocabularies using shared and agreed-upon glossaries and dictionaries from different domains as a departure point, and then to complete evolutive processes aimed at semantic extension to create ontologies, which could then be com- bined with other ontologies used in information systems running in both conventional and digital libraries for index- ing as well as for supporting document searches. there are examples of glossaries that have been transformed into ontologies, such as the cambridge healthtech institute’s “pharmaceutical ontologies glossary and taxonomy” (http://www.genomicglossaries.com/content/ontolo gies.asp), which is an “evolving terminology for emerging technologies.” ■■ ieee standard glossary of software engineering terminology to demonstrate our proposed method, we will use a real glossary belonging to the computer science field, although it is possible to use any other. the glossary, available in electronic format (pdf), defines approxi- mately , terms in the domain of software engineering (figure ). topics include addressing assembling, compil- ing, linking, loading; computer performance evaluation; figure . cover of the glossary document generating collaborative systems for digital libraries | hilera et al. . define the classes and the class hierarchy . define the properties of classes (slots) . define the facets of the slots . create instances as outlined in the introduction, the ontology devel- oped using our method is a terminological one. therefore we can ignore the first two steps in noy’s and mcguinness’ process as the concepts of the ontology coincide with the terms of the glossary used. any ontology development process must take into account the basic stages of the life cycle, but the way of organizing the stages can be different in different meth- ods. in our case, since the ontology has a terminological character, we have established an incremental develop- ment process that supposes the natural evolution of the glossary from its original format (dictionary or vocabu- lary format) into an ontology. the proposed life cycle establishes a series of steps or phases that will result in intermediate knowledge representation tools, with the final product, the ontology, being the most semantically rich (figure ). therefore this is a product-driven process, in which the aim of every step is to obtain an intermediate product useful on its own. the intermediate products and the final examples associated with the described concept. in the resulting ontology, the examples were included as instances of the corresponding class. in figure , it can be seen that the definition refers to another glossary on programming languages (std . ), which is a part of the series of dic- tionaries related to computer science (“ieee std ,” figure ). other glossaries which are men- tioned in relation to some references about term definitions are . , . , . , . , and . . to avoid redundant definitions and pos- sible inconsistencies, links must be implemented between ontologies developed from those glossa- ries that include common concepts. the ontology generation process presented in this paper is meant to allow for integration with other ontolo- gies that will be developed in the future from the other glossaries. in addition to the explicit references to other terms within the glossary and to terms from other glos- saries, the textual definition of a concept also has implicit references to other terms. for example, from the phrase “provides features designed to facilitate expression of data structures” included in the definition of the term high order language (figure ), it is possible to determine that there is an implicit relationship between this term and the term data structure, also included in the glossary. these relationships have been considered in establishing the properties of the concepts in the developed ontology. ■■ ontology development process many ontology development methods presuppose a life cycle and suggest technologies to apply during the pro- cess of developing an ontology. the method described by noy and mcguinness is helpful when beginning this process for the first time. they establish a seven-step process: . determine the domain and scope of the ontology . consider reusing existing ontologies . enumerate important terms in the ontology figure . example of term definition in the ieee glossary figure . ieee computer science glossaries —standard dictionary of computer terminology . —standard glossary of mathematics of computing terminology . —standard glossary of computer applications terminology . —standard glossary of modeling and simulation terminology . —standard glossary of image processing terminology . —standard glossary of data management terminology . —standard glossary of computer graphics terminology . —standard glossary of computer networking terminology . —standard glossary of artificial intelligence terminology . —standard glossary of computer security and privacy terminology . —standard glossary of computer hardware terminology . —standard glossary of theory of computation terminology . —standard glossary of software engineering terminology . —standard glossary of computer languages terminology high order language (hol). a programming language that requires little knowledge of the computer on which a program will run, can be translated into several difference machine languages, allows symbolic naming of operations and addresses, provides features designed to facilitate expression of data structures and program logic, and usually results in several machine instructions for each program state- ment. examples include ada, cobol, fortran, algol, pascal. syn: high level language; higher order language; third gen- eration language. contrast with: assembly language; fifth generation language; fourth generation language; machine language. note: specific languages are defined in p . information technology and libraries | december since there are terms with different meanings (up to five in some cases) in the ieee glossary of software engineering terminology, during dictionary development we decided to create different concepts (classes) for the same term, associating a number to these concepts to differentiate them. for example, there are five different definitions for the term test, which is why there are five concepts (test –test ), corresponding to the five meanings of the term: ( ) an activity in which a system or compo- nent is executed under specified conditions, the results are observed or recorded, and an evaluation is made of some aspect of the system or component; ( ) to conduct an activity as in ( ); ( ) a set of one or more test cases; ( ) a set of one or more test procedures; ( ) a set of one or more test cases and procedures. taxonomy the proposed lifecycle establishes a stage for the con- version of a dictionary into a taxonomy, understanding taxonomy as an instrument of concepts categorization, product are a dictionary, which has a formal and computer processed structure, with the terms and their definitions in xml format; a taxonomy, which reflects the hierarchic rela- tionships between the terms; a thesaurus, which includes other relationships between the terms (for example, the synonymy relationship); and, finally, the ontology, which will include the hierarchy, the basic relationships of the the- saurus, new and more complex semantic relationships, and restrictions in form of axioms expressed using description logics. the following paragraphs describe the way each of these products is obtained. dictionary the first step of the proposed development process con- sists of the creation of a dictionary in xml format with all the terms included in the ieee standard glossary of software engineering terminology and their related defini- tions. this activity is particularly mechanical and does not need human intervention as it is basically a transfor- mation of the glossary from its original format (pdf) into a format better suited to the development process. all formats considered for the dictionary are based on xml, and specifically on rdf and rdf schema. in the end, we decided to work with the standards daml+oil and owl, though we are not opposed to working with other languages, such as skos or xmi, in the future. (in the latter case, it would be possible to model the intermediate products and the ontology in uml graphic models stored in xml files.) in our project, the design and implementation of all products has been made using an ontology editor. we have used oiled (with oilviz plugin) as editor, both because of its simplicity and because it allows the exportation to owl and daml formats. however, with future maintenance and testing in mind, we decided to use protégé (with owl plugin) in the last step of the process, because this is a more flexible environment with extensible mod- ules that integrate more functionality such as ontology annotation, evaluation, middleware service, query and inference, etc. figure shows the dictionary entry for “high order language,” which appears in figure . note that the dic- tionary includes only owl:class (or daml:class) to mark the term; rdf:label to indicate the term name; and rdf:comment to provide the definition included in the original glossary. figure . ontology development process highorderlanguage figure . example of dictionary entry generating collaborative systems for digital libraries | hilera et al. example, when analyzing the definition of the term com- piler: “(is) a computer program that translates programs expressed in a high order language into their machine language equivalent,” it is possible to deduce that com- piler is a subconcept of computer program, which is also included in the glossary.) in addition to the lexical or syn- tactic analysis, it is necessary for an expert in the domain to perform a semantic analysis to complete the develop- ment of the taxonomy. the implementation of the hierarchical relation- ships among the concepts is made using rdfs:subclassof, regardless of whether the taxonomy is implemented in owl or daml format, since both languages specify this type of relationship in the same way. figure shows an example of a hierarchical relationship included in the definition of the concept pictured in figure . thesaurus according to the international organization for standardization (iso), a thesaurus is “the vocabulary of a controlled indexing language, formally organized in order to make explicit the a priori relations between concepts (for example ‘broader’ and ‘narrower’).” this definition establishes the lexical units and the semantic relationships between these units as the elements that constitute a the- saurus. the following is a sample of the lexical units: ■■ descriptors (also called “preferred terms”): the terms used consistently when indexing to represent a con- cept that can be in documents or in queries to these documents. the iso standard introduces the option of adding a definition or an application note to every term to establish explicitly the chosen meaning. this note is identified by the abbreviation sn (scope note), as shown in figure . ■■ non-descriptors (“non-preferred terms”): the syn- onyms or quasi-synonyms of a preferred term. a nonpreferred term is not assigned to documents submitted to an indexing process, but is provided as an entry point in a thesaurus to point to the appropri- ate descriptor. usually the descriptors are written in capital letters and the nondescriptors in small letters. ■■ compound descriptors: the terms used to represent complex concepts and groups of descriptors, which allow for the structuring of large numbers of thesau- rus descriptors into subsets called micro-thesauri. in addition to lexical units, other fundamental elements of a thesaurus are semantic relationships between these units. the more common relationships between lexical units are the following: ■■ equivalence: the relationship between the descrip- tors and the nondescriptors (synonymous and that is, as a systematical classification in a traditional way. as gilchrist states, there is no consensus on the meaning of terms like taxonomy, thesaurus, or ontology. in addi- tion, much work in the field of ontologies has been done without taking advantage of similar work performed in the fields of linguistics and library science. this situa- tion is changing because of the increasing publication of works that relate the development of ontologies to the development of “classic” terminological tools (vocabular- ies, taxonomies, and thesauri). this paper emphasizes the importance and useful- ness of the intermediate products created at each stage of the evolutive process from glossary to ontology. the end product of the initial stage is a dictionary expressed as xml. the next stage in the evolutive process (figure ) is the transformation of that dictionary into a tax- onomy through the addition of hierarchical relationships between concepts. to do this, it is necessary to undertake a lexical- semantic analysis of the original glossary. this can be done in a semiautomatic way by applying natural language processing (nlp) techniques, such as those recommended by morales-del-castillo et al., for creat- ing thesauri. the basic processing sequence in linguistic engineering comprises the following steps: ( ) incorpo- rate the original documents (in our case the dictionary obtained in the previous stage) into the information sys- tem; ( ) identify the language in which they are written, distinguishing independent words; ( ) “understand” the processed material at the appropriate level; ( ) use this understanding to transform, search, or traduce data; ( ) produce the new media required to present the produced outcomes; and finally, ( ) present the final outcome to human users by means of the most appropriate periph- eral device—screen, speakers, printer, etc. an important aspect of this process is natural lan- guage comprehension. for that reason, several different kinds of programs are employed, including lemmatizers (which implement stemming algorithms to extract the lexeme or root of a word), morphologic analyzers (which glean sentence information from their constituent ele- ments: morphemes, words, and parts of speech), syntactic analyzers (which group sentence constituents to extract elements larger than words), and semantic models (which represent language semantics in terms of concepts and their relations, using abstraction, logical reasoning, orga- nization and data structuring capabilities). from the information in the software engineering dictionary and from a lexical analysis of it, it is possible to determine a hierarchical relationship when the name of a term contains the name of another one (for example, the term language and the terms programming language and hardware design language), or when expressions such as “is a” linked to the name of another term included in the glossary appear in the text of the term definition. (for information technology and libraries | december indicating that high order language relates to both assembly and machine languages. the life cycle proposed in this paper (figure ) includes a third step or phase that transforms the taxonomy obtained in the previous phase into a thesaurus through the incorporation of relationships between the concepts that complement the hierarchical relations included in the taxonomy. basically, we have to add two types of relation- ships—equivalence and associative, represented in the standard thesauri with uf (and use) and rt respectively. we will continue using xml to implement this new product. there are different ways of implementing a thesaurus using a language based on xml. for example, matthews et al. proposed a standard rdf format, where as hall created an ontology in daml. in both cases, the authors modeled the general structure of quasi-synonymous). iso establishes that the abbrevia- tion uf (used for) precedes the nondescriptors linked to a descriptor; and the abbreviation use is used in the opposite case. for example, a thesaurus developed from the ieee glossary might include a descriptor “high order language” and an equivalence relationship with a nondescriptor “high level language” (figure ). ■■ hierarchical: a relationship between two descrip- tors. in the thesaurus one of these descriptors has been defined as superior to the other one. there are no hierarchical relationships between nondescrip- tors, nor between nondescriptors and descriptors. a descriptor can have no lower descriptors or several of them, and no higher descriptors or several of them. according to the iso standard, hierarchy is expressed by means of the abbreviations bt (broader term), to indicate the generic or higher descriptors, and nt (narrower term), to indicate the specific or lower descriptors. the term at the head of the hierarchy to which a term belongs can be included, using the abbreviation tt (top term). figure presents these hierarchical relationships. ■■ associative: a reciprocal relationship that is estab- lished between terms that are neither equivalent nor hierarchical, but are semantically or conceptually associated to such an extent that the link between them should be made explicit in the controlled vocabulary on the grounds that it may suggest additional terms for use in indexing or retrieval. it is generally indicated by the abbreviation rt (related term). there are no associative relationships between nondescriptors and descriptors, or between descriptors already linked by a hierarchical relation. it is possible to establish associative relationships between descriptors belonging to the same or differ- ent category. the associative relationships can be of very different types. for example, they can represent causality, instrumentation, location, similarity, origin, action, etc. figure shows two associative relations, .. high order language (descriptor) sn a programming language that... uf high level language (no-descriptor) uf third generation language (no-descriptor) tt language bt programming language nt object oriented language nt declarative language rt assembly language (contrast with) rt machine language (contrast with) .. high level language use high order language .. third generation language use high order language .. figure . fragment of a thesaurus entry figure . example of taxonomy entry ... generating collaborative systems for digital libraries | hilera et al. terms. for example: . or using the glossary notation: . ■■ the rest of the associative relationships (rt) that were included in the thesaurus correspond to the cross-references of the type “contrast with” and “see also” that appear explicitly in the ieee glossary. ■■ neither compound descriptors nor groups of descrip- tors have been implemented because there is no such structure in the glossary. ontology ding and foo state that “ontology promotes standard- ization and reusability of information representation through identifying common and shared knowledge. ontology adds values to traditional thesauri through deeper semantics in digital objects, both conceptually, relationally and machine understandably.” this seman- tic richness may imply deeper hierarchical levels, richer relationships between concepts, the definition of axioms or inference rules, etc. the final stage of the evolutive process is the transfor- mation of the thesaurus created in the previous stage into an ontology. this is achieved through the addition of one or more of the basic elements of semantic complexity that differentiates ontologies from other knowledge represen- tation standards (such as dictionaries, taxonomies, and thesauri). for example: ■■ semantic relationships between the concepts (classes) of the thesaurus have been added as properties or ontology slots. ■■ axioms of classes and axioms of properties. these are restriction rules that are declared to be sat- isfied by elements of ontology. for example, to establish disjunctive classes ( ), have been defined, and quantification restrictions (existential or universal) and cardinality restrictions in the relation- ships have been implemented as properties. software based on techniques of linguistic analysis has been developed to facilitate the establishment of the properties and restrictions. this software analyzes the definition text for each of the more than , glossary terms (in thesaurus format), isolating those words that a thesaurus from classes (rdf:class or daml:class) and properties (rdf:property or daml:objectproperty). in the first case they proposed five classes: thesaurusobject, concept, topconcept, term, scopenote; and several properties to implement the relations, like hasscope- note (sn), isindicatedby, preferredterm, usedfor (uf), conceptrelation, broaderconcept (bt), narrowerconcept (nt), topofhierarchy (tt) and isrelatedto (rt). recently the w c has developed the skos specifica- tion, created to define knowledge organization schemes. in the case of thesauri, skos includes specific tags, such as skos:concept, skos:scopenote (sn), skos:broader (bt), skos:narrower (nt), skos:related (rt), etc., that are equivalent to those listed in the previous paragraph. our specification does not make any statement about the formal relationship between the class of skos concept schemes and the class of owl ontologies, which will allow different design patterns to be explored for using skos in combination with owl. although any of the above-mentioned formats could be used to implement the thesaurus, given that the end- product of our process is to be an ontology, our proposal is that the product to be generated during this phase should have a format compatible with the final ontology and with the previous taxonomy. therefore a minimal number of changes will be carried out on the product created in the previous step, resulting in a knowledge representation tool similar to a thesaurus. that tool does not need to be modified during the following (final) phase of transformation into an ontology. nevertheless, if for some reason it is necessary to have the thesaurus in one of the other formats (such as skos), it is possible to apply a simple xslt transformation to the product. another option would be to integrate a thesaurus ontology, such as the one proposed by hall, with the ontology represent- ing the ieee glossary. in the thesaurus implementation carried out in our project, the following limitations have been considered: ■■ only the hierarchical relationships implemented in the taxonomy have been considered. these include relationsips of type “is-a,” that is, generalization rela- tionships or type–subset relationships. relationships that can be included in the thesaurus marked with tt, bt, and nt, like relations of type “part of” (that is, partative relationships) have not been considered. instead of considering them as hierarchical relation- ships, the final ontology includes the possibility of describing classes as a union of classes. ■■ the relationships of synonymy (uf and use) used to model the cross-references in the ieee glossary (“syn” and “see,” respectively) were implemented as equiv- alent terms, that is, as equivalent axioms between classes (owl:equivalentclass or daml:sameclassas), with inverse properties to reflect the preference of the information technology and libraries | december match the name of other glossary terms (or a word in the definition text of other glossary terms). the isolated words will then be candidates for a relationship between both of them. (figure shows the candidate properties obtained from the software engineering glossary.) the user then has the option of creating relationships with the identified candidate words. the user must indicate, for every relationship to be created, the restriction type that it represents as well as existential or universal quan- tification or cardinality (minimum or maximum). after confirming this information, the program updates the file containing the ontology (owl or daml), adding the property to the class that represents the processed term. figure shows an example of the definition of two prop- erties and its application to the class highorderlanguage: a property express with existential quantification over the class datastructure to indicate that a language must repre- sent at least one data structure; and a property translateto of universal type to indicate that any high-level language is translated into machine language (machinelanguage). ■■ results, conclusions, and future work the existence of ontologies of specific knowledge domains (software engineering in this case) facilitates the process of finding resources about this discipline on the semantic web and in digital libraries, as well as the reuse of learn- ing objects of the same domain stored in repositories available on the web. when a new resource is indexed in a library catalog, a new record that conforms to the ontology conceptual data model may be included. it will be necessary to assign its properties according to the concept definition included in the ontology. the user may later execute semantic queries that will be run by the search system that will traverse the ontology to identify the concept in which the user was interested to launch a wider query including the resources indexed under the concept. ontologies, like the one that has been “evolved,” may also be used in an open way to index and search for resources on the web. in that case, however, semantic search engines such as swoogle (http://swoogle.umbc .edu/), are required in place of traditional syntactic search engines, such as google. the creation of a complete ontology of a knowledge domain is a complex task. in the case of the domain presented in this paper, that of software engineering, although there have been initiatives toward ontology cre- ation that have yielded publications by renowned authors in the field, a complete ontology has yet to be created and published. this paper has described a process for developing a modest but complete ontology from a glossary of ter- minology, both in owl format and daml+oil format, accept access accomplish account achieve adapt add adjust advance affect aggregate aid allocate allow allow symbolic naming alter analyze apply approach approve arrangement arrive assign assigned by assume avoid await begin break bring broke down builds call called by can be can be input can be used as can operate in cannot be usedas carry out cause change characterize combine communicate compare comply comprise conduct conform consist constrain construct contain contains no contribute control convert copy correct correspond count create debugs decompiles decomposedinto decrease define degree delineate denote depend depict describe design designate detect determine develop development direct disable disassembles display distribute divide document employ enable encapsulate encounter ensure enter establish estimate establish evaluate examine exchange execute after execute in executes expand express express as extract facilitate fetch fill follow fulfil generate give give partial given constrain govern have have associated have met have no hold identify identify request ignore implement imply improve incapacitate include incorporate increase indicate inform initiate insert install intend interact with interprets interrelate investigate invokes is is a defect in is a form of is a method of is a mode of is a part is a part of is a sequence is a sequenceof is a technique is a techniqueof is a type is a type of is ability is activated by is adjusted by is applied to is based is called by is composed is contained is contained in is establish is established is executed after is executed by is incorrect is independent of is manifest is measured in is not is not subdivided in is part is part of is performed by is performed on is portion is process by is produce by is produce in is ratio is represented by is the output is the result of is translated by is type is used is used in isolate know link list load locate maintain make make up may be measure meet mix modify monitors move no contain no execute no relate no use not be connected not erase not fill not have not involve not involving not translate not use occur occur in occur in a operate operatewith optimize order output parses pas pass test perform permit permitexecute permit the execution pertaining place preclude predict prepare prescribe present present for prevent preventaccessto process produce produce no propose provide rank reads realize receive reconstruct records recovery refine reflect reformat relate relation release relocates remove repair replace represent request require reserve reside restore restructure result resume retain retest returncontrolto reviews satisfy schedule send server set share show shutdown specify store store in structure submission of supervise supports suppress suspend swap synchronize take terminate test there are no through throughout transfer transform translate transmit treat through understand update use use in use to utilize value verify work in writes figure . candidate properties obtained from the linguistic analysis of the software engineering glossary generating collaborative systems for digital libraries | hilera et al. to each term.) we defined properties or relationships between these classes. these are based on a semiauto- mated linguistic analysis of the glossary content (for example, allow, convert, execute, operatewith, produces, translate, transform, utilize, workin, etc.), which will be refined in future versions. the authors’ aim is to use this ontology, which we have called ontoglose (ontology glossary software engineering), to unify the vocabulary. ontoglose will be used in a more ambitious project, whose purpose is the development of a complete ontology in software engi- neering from the swebok guide. although this paper has focused on this ontology, the method that has been described may be used to generate an ontology from any dictionary. the flexibility that owl permits for ontology description, along with its compat- ibility with other rdf-based metadata languages, makes possible interoperability between ontologies and between ontologies and other controlled vocabularies and allows for the building of merged representations of multiple knowledge domains. these representations may eventu- ally be used in libraries and repositories to index and search for any kind of resource, not only those related to the original field. ■■ acknowledgments this research is co-funded by the spanish ministry of industry, tourism and commerce profit program (grant tsi- - - ). the authors also want to acknowledge support from the tifyc research group at the university of alcala. references and notes . m. dörr et al., state of the art in content standards (amster- dam: ontoweb consortium, ). . d. soergel, “the rise of ontologies or the reinvention of classification,” journal of the american society for information science , no. ( ): – ; a. gilchrist, “thesauri, tax- onomies and ontologies—an etymological note,” journal of documentation , no. ( ): – . . b. j. wielinga et al., “from thesaurus to ontology,” pro- ceedings of the st international conference on knowledge capture (new york: acm, ): – : j. qin and s. paling, “con- verting a controlled vocabulary into an ontology: the case of gem,” information research ( ): . . according to van heijst, schereiber, and wielinga, ontolo- gies can be classified as terminological ontologies, information ontologies, and knowledge modeling ontologies; terminological ontologies specify the terms that are used to represent knowl- edge in the domain of discourse, and they are in use principally to unify vocabulary in a certain domain. g. van heijst, a. t. which is ready to use in the semantic web. as described at the opening of this article, our aim has been to create a lightweight ontology as a first version, which will later be improved by including more axioms and relationships that increase its semantic expressiveness. we have tried to make this first version as tailored as possible to the initial glossary, knowing that later versions will be improved by others who might take on the work. such improvements will increase the ontology’s utility, but will make it a less- faithful representation of the ieee glossary from which it was derived. the ontology we have developed includes , classes that correspond to the same number of concepts represented in the ieee glossary. (included in this num- ber are the different meanings that the glossary assigns ... figure . example of ontology entry information technology and libraries | december . w c, skos; object management group, xml metadata interchange (xmi), , http://www.omg.org/technology/doc- uments/formal/xmi.htm (accessed oct. , ). . uml (unified modeling language) is a standardized general-purpose modeling language (http://www.uml.org). nowadays, different uml plugins for ontologies’ editors exist. these plugins allow working with uml graphic models. also, it is possible to realize the uml models with a case tool, to export them to xml format, and to transform them to the ontol- ogy format (for example, owl) using a xslt sheet, as the one published in d. gasevic, “umltoowl: converter from uml to owl,” http://www.sfu.ca/~dgasevic/projects/umltoowl/ (accessed oct. , ). . gilchrist, “thesauri, taxonomies and ontologies.” . soergel, “the rise of ontologies or the reinvention of classification.” . j. m. morales-del-castillo et al., “a semantic model of selective dissemination of information for digital libraries,” information technology & libraries , no. ( ): – . . international standards organization, iso : doc- umentation—guidelines for the establishment and develop- ment of monolingual thesauri (geneve: international standards organization, ). . b. m. matthews, k. miller, and m. d. wilson, “a thesau- rus interchange format in rdf,” , http://www.w c.rl.ac .uk/swad/thes_links.htm (accessed feb. , ). . m. hall, “call thesaurus ontology in daml,” dynam- ics research corporation, , http://orlando.drc.com/daml/ ontology/call-thesaurus (accessed oct. , ). . ibid. . y. ding and s. foo, “ontology research and develop- ment. part —a review of ontology generation,” journal of information science , no. ( ): – . see also b. h. kwas- nik, “the role of classification in knowledge representation and discover,” library trends ( ): – . . s. otón et al., “service oriented architecture for the imple- mentation of distributed repositories of learning objects,” international journal of innovative computing, information & con- trol ( ), forthcoming. . o. mendes and a. abran, “software engineering ontol- ogy: a development methodology,” metrics news ( ): – ; c. calero, f. ruiz, and m. piattini, ontologies for software engineering and software technology (berlin: springer, ). . ieee, guide to the software engineering body of knowledge (swebok) (los alamitos, calif.: ieee computer society, ), http:// www.swebok.org (accessed oct. , ). schereiber, and b. j. wielinga, “using explicit ontologies in kbs development,” international journal of human & computer studies , no. / ( ): – . . r. neches et al., “enabling technology for knowledge sharing,” ai magazine , no. ( ): – . . o. corcho, f. fernández-lópez, and a. gómez-pérez, “methodologies, tools and languages for buildings ontologies. where is their meeting point?” data & knowledge engineering , no. ( ): – . . intitute of electrical and electronics engineers (ieee), ieee std . - (r ): ieee standard glossary of software engineering terminology (reaffirmed ) (new york: ieee, ). . j. krause, “semantic heterogeneity: comparing new semantic web approaches with those of digital libraries,” library review , no. ( ): – . . t. berners-lee, j. hendler, and o. lassila, “the semantic web,” scientific american , no. ( ): – . . world wide web consortium (w c), resource description framework (rdf): concepts and abstract syntax, w c recommen- dation february , http://www.w .org/tr/rdf-concepts/ (accessed oct. , ). . world wide web consortium (w c), web ontology lan- guage (owl), , http://www.w .org/ /owl (accessed oct. , ). . world wide web consortium (w c), skos simple knowledge organization system, , http://www.w .org/ tr/ /rec-skos-reference- / (accessed oct. , ). . m. m. yee, “can bibliographic data be put directly onto the semantic web?” information technology & libraries , no. ( ): - . . l. f. spiteri, “the structure and form of folksonomy tags: the road to the public library catalog,” information technology & libraries , no. ( ): – . . corcho, fernández-lópez, and gómez-pérez, “method- ologies, tools and languages for buildings ontologies.” . ieee, ieee std . - (r ). . n. f. noy and d. l. mcguinness, “ontology develop- ment : a guide to creating your first ontology,” , stan- ford university, http://www-ksl.stanford.edu/people/dlm/ papers/ontology-tutorial-noy-mcguinness.pdf (accessed sept , ). . d. baader et al., the description logic handbook (cam- bridge: cambridge univ. pr., ). . world wide web consortium, daml+oil reference description, , http://www.w .org/tr/daml+oil-reference (accessed oct. , ); w c, owl. bridging the gap: self-directed staff technology training | quinney, smith, and galbraith kayla l. quinney, sara d. smith, and quinn galbraith bridging the gap: self-directed staff technology training of hbll patrons. as anticipated, results indicated that students frequently use text messages, social networks, blogs, etc., while fewer staff members use these technolo- gies. for example, percent of the students reported that they write a blog, while only percent of staff and fac- ulty do so. also, percent of the students and only percent of staff and faculty indicated that they belonged to a social network. after concluding that staff and faculty were not as connected as their student patrons are to tech- nology, library administration developed the technology challenge to help close this gap. the technology challenge was a self-directed training program requiring participants to explore new technol- ogy on their own by spending at least fifteen minutes each day learning new technology skills. this program was successful in promoting lifelong learning by teach- ing technology applicable to the work and home lives of hbll employees. we will first discuss literature that shows how technology training can help academic librar- ians connect with student patrons, and then we will describe the technology challenge and demonstrate how it aligns with the principles of self-directed learning. the training will be evaluated by an analysis of the results of two surveys given to participants before and after the technology challenge was implemented. ■■ library . and “librarian . ” hbll wasn’t the first to notice the gap between librar- ians and students, mcdonald and thomas noted that “gaps have materialized,” and library technology does not always “provide certain services, resources, or possibilities expected by emerging user populations like the millennial generation.” college students, who grew up with technol- ogy, are “digital natives,” while librarians, many having learned technology later in life, are “digital immigrants.” the “digital natives” belong to the millennial generation, described by shish and allen as a generation of “learners raised on and confirmed experts in the latest, fastest, cool- est, greatest, newest electronic technologies.” according to sweeny, when students use libraries, they expect the same “flexibility, geographic independence, speed of response, time shifting, interactivity, multitasking, and time savings” provided by the technology they use daily. students are undergraduates, as members of the millennial generation, are proficient in web . technology and expect to apply these technologies to their coursework—including schol- arly research. to remain relevant, academic libraries need to provide the technology that student patrons expect, and academic librarians need to learn and use these tech- nologies themselves. because leaders at the harold b. lee library of brigham young university (hbll) perceived a gap in technology use between students and their staff and faculty, they developed and implemented the technology challenge, a self-directed technology training program that rewarded employees for exploring technology daily. the purpose of this paper is to examine the technology challenge through an analysis of results of surveys given to participants before and after the technology challenge was implemented. the program will also be evaluated in terms of the adult learning theories of andragogy and self- directed learning. hbll found that a self-directed approach fosters technology skills that librarians need to best serve students. in addition, it promotes lifelong learning hab- its to keep abreast of emerging technologies. this paper offers some insights and methods that could be applied in other libraries, the most valuable of which is the use of self-directed and andragogical training methods to help academic libraries better integrate modern technologies. l eaders at the harold b. lee library of brigham young university (hbll) began to suspect a need for technology training when employees were asked during a meeting if they owned an ipod or mp player. out of the twenty attendees, only two raised their hands—one of whom worked for it. perceiving a technol- ogy gap between hbll employees and student patrons, library leaders began investigating how they could help faculty and staff become more proficient with the tech- nologies that student patrons use daily. to best serve student patrons, academic librarians need to be proficient with the technologies that student patrons expect. hbll found that a self-directed learning approach to staff tech- nology training not only fosters technology skills, but also promotes lifelong learning habits. to further examine the technology gap between librar- ians and students, the hbll staff, faculty, and student employees were given a survey designed to explore generational differences in media and technology use. student employees were surveyed as representatives of the larger student body, which composes the majority kayla l. quinney (quinster @gmail.com) is research spe- cialist, sara d. smith (saradsmith@gmail.com) is research specialist, and quinn galbraith (quinn_galbraith@byu.edu) is library human resource training and development manager, brigham young university library, provo, utah. information technology and libraries | december . ,” a program that “focuses on self-exploration and encourages staff to learn about new technologies on their own.” learning . encouraged library staff to explore web . tools by completing twenty-three exercises involving new technologies. plcmc’s program has been replicated by more than libraries and organizations worldwide, and several libraries have written about their experiences, including academic and public libraries. these programs—and the technology challenge implemented by hbll—integrate the theories of adult learning. in the s and s, malcolm knowles intro- duced the theory of andragogy to describe the way adults learn. knowles described adults as learners who ( ) are self-directed, ( ) use their experiences as a resource for learning, ( ) learn more readily when they experience a need to know, ( ) seek immediate application of knowl- edge, and ( ) are best motivated by internal rather than external factors. the theory and practice of self-directed learning grew out of the first learning characteristic and assumes that adults prefer self-direction in determining and achieving learning goals, and therefore learners exer- cise independence in determining how and what they learn. these theories have had a considerable effect on adult education practice and employee development programs. when adults participate in trainings that align with the assumptions of andragogy, they are more likely to retain and apply what they have learned. ■■ the technology challenge hbll’s technology challenge is similar to learning . in that it encourages self-directed exploration of web . technologies, but it differs in that participants were even more self-directed in exploration and that they were asked to participate daily. these features encouraged more self-directed learning in areas of participant interest as well as habit formation. it is not our purpose to critique learning . , but to provide some evidence and analysis to demonstrate the success of hands-on, self-directed training approaches and to suggest other ways for librar- ies to apply self-directed learning to technology training. the technology challenge was implemented from june to january . hbll staff included full-time employees, of whom participated in the challenge. (the student employees were not involved.) participants were asked to spend fifteen minutes each day learning a new technology skill. hbll leaders used rewards to make the program enjoyable and to motivate participation: for each minute spent learning technology, participants earned one point, and when one thousand points were earned, the participant would receive a gift certificate to the campus bookstore. staff and faculty participated and tracked their progress through an online masters of “informal learning”; that is, they are accus- tomed to easily and quickly gathering information relevant to their lives from the internet and from friends. shish and allen claimed that millennials prefer “interactive, hyper-linked multimedia over the traditional static, text- oriented printed items. they want a sense of control; they need experiential and collaborative approaches rather than formal, librarian-guided, library-centric services.” these students arrive on campus expecting “to handle the chal- lenges of scholarly research” using similar methods and technologies. interactive technologies such as blogs, wikis, streaming media applications, and social networks, are referred to as “web . .” abram argued that web . technology “could be useful in an enterprise, institutional research, or com- munity environment, and could be driven or introduced by the library.” “library . ” is a concept referring to a library’s integration of these technologies; it is essentially the use of “web . opportunities in a library environ- ment.” manesss described library . is user-centered, social, innovative, and provider of a multimedia experi- ences. it is a community that “blurs the line between librarian and patron, creator and consumer, authority and novice.” libraries have been using web . tech- nology such as blogs, wikis, and social networks to better serve and connect with patrons. blogs allow librar- ies to “provide news, information and links to internet resources,” and wikis create online study groups and “build a shared knowledge repository.” social networks can be particularly useful in connecting with undergradu- ate students: millennials use technology to collaborate and make collective decisions, and libraries can capitalize on this tendency by using social networks, which for students would mean, as bates argues, “an informational equiva- lent of the reliance on one’s facebook friends.” students expect library . —and as libraries integrate new technologies, the staff and faculty of academic librar- ies need to become “librarian . .” according to abram, librarian . understands users and their needs “in terms of their goals and aspirations, workflows, social and con- tent needs, and more. librarian . is where the user is, when the user is there.” the modern library user “needs the experience of the web . . . to learn and succeed,” and the modern librarian can help patrons transfer technology skills to information seeking. librarian . is prepared to help patrons familiar with web . to “lever- age these [technologies] to make a difference in reaching their goals.” therefore staff and faculty “must become adept at key learning technologies themselves.” stephen abram asked, “are the expectations of our users increas- ing faster than our ability to adapt?” and this same concern motivated hbll and other institutions to initiate staff technology training programs. the public library of charlotte and mecklenburg county of north carolina (plcmc) developed “learning bridging the gap: self-directed staff technology training | quinney, smith, and galbraith their ability to learn and use technology. to be eligible to receive the gift card, participants were required to take this exit survey. sixty-four participants, all of whom had met or exceeded the thousand-point goal, chose to complete this survey, so the results of this survey repre- sent the experiences of percent of the participants. of course, if those who had not completed the technology challenge had taken the survey the results may have been different, but the results do show how those who chose to actively participate reacted to this training program. the survey included both quantifiable and open-ended questions (see appendix b for survey results and a list of the open-ended questions). the survey results, along with an analysis of the structure of the challenge itself, demonstrates that the program aligns with knowles’s five principles of andragogy to successfully help employees develop both technology skills and learning habits. self-direction the technology challenge was self-directed because it gave participants the flexibility to select which tasks and challenges they would complete. garrison wrote that in a self-directed program, “learners should be provided with choices of how they wish to proactively carry out the learning process. material resources should be available, approaches suggested, flexible pacing accommodated, and questioning and feedback provided when needed.” hbll provided a variety of challenges and training sessions related to various technologies. technology challenge participants were given the independence to choose which learning methods to use, including which training sessions to attend and which challenges to complete. according to the exit survey, the most popular training methods were small, instructor-led groups, followed by self-learning through reading books and articles. group training sessions were organized by hbll leadership and addressed topics such as microsoft office, rss feeds, computer organization skills, and multimedia software. other learning methods included web tutorials, dvds, large group discussions, and one-on-one tutoring. the group training classes preferred by hbll employees may be considered more teacher-directed than self-directed, but the technology challenge was self-directed as a whole in that learners were given the opportunity to choose what they learned and how they learned it. the structure of the technology challenge allowed participants to set their own pace. staff and faculty were given several months to complete the challenge and were responsible to pace themselves. on the exit survey, one participant commented: “if i didn’t get anything done one week, there wasn’t any pressure.” another enjoyed flexibility in deciding when and where to complete the tasks: “i liked being able to do the challenge anywhere. when i had a few minutes between appointments, classes, board game called “techopoly.” participation was voluntary, and staff and faculty were free to choose which tasks and challenges they would complete. tasks fell into one of four categories: software, hardware, library technology, and the internet. participants were required to complete one hundred points in each category, but beyond that, were able to decide how to spend their time. examples of tasks included attending workshops, exploring online tutori- als, and reading books or articles about a relevant topic. for each hundred points earned, participants could com- plete a mini-challenge, which included reading blogs or e-books, listening to podcasts, or creating a photo cd (see appendix a for a more complete list). participants who completed fifteen out of twenty possible challenges were entered into a drawing for another gift certificate. before beginning the challenge, all participants were surveyed about their current use of technology. on this survey, they indicated that they were most uncomfortable with blogs, wikis, image editors, and music players. these results provided a focus for technology challenge trainings and mini-challenges. while not all of these technologies may apply directly to their jobs, percent indicated that they were interested in learning them. forty-four percent reported that time was the greatest impediment to learn- ing new technology; therefore the daily fifteen-minute requirement was introduced with the hope that it was small enough to be a good incentive to participate but substantial enough to promote habit formation and allow employees enough time to familiarize themselves with the technology. although some productivity may have been lost due to the time requirement (especially in cases where participants may have spent more than the required time), library leaders felt that technology training was an investment in hbll employees and that, at least for a few months, it was worth any potential loss in productiv- ity. because participants could chose how and when they learned technology, they could incorporate the challenge into their work schedules according to their own needs, interests, and time constraints. of ninety-six participants, sixty-six reached or exceeded the thousand-point goal, and eight participants earned more than two thousand points. ten participants earned between five hundred and one thousand points, and another six earned between one hundred and five hundred. although not all participants completed the challenge, most were involved to some extent in learning technology during this time. ■■ the technology challenge and adult learning after finishing the challenge, participants took an exit survey to evaluate the experience and report changes in information technology and libraries | december were willing, even excited, to learn technology skills: percent “agreed” and percent “strongly agreed” that they were interested in learning new technology. their desire to learn was cultivated by the survey itself, which helped them recognize and focus on this interest, and the challenge provided a way for employees to channel their desire to learn technology. immediate application learners need to see an opportunity for immediate application of their knowledge: ota et al. explained that “they want to learn what will help them perform tasks or deal with problems they confront in everyday situations and those presented in the context of application to real life.” because of the need for immediate application, the technology challenge encouraged staff and faculty to learn technology skills directly related to their jobs—as well as technology that is applicable to their personal or home lives. hbll leaders hoped that as staff became more comfortable with technology in general, they would be motivated to incorporate more complex technologies into their work. here is one example of how the technology challenge catered to adult learners’ need to apply what they learn: before designing the challenge, hbll held a training session to teach employees the basics of photoshop. even though attendees were on the clock, the turnout was discouraging. library leaders knew they needed to try something new. in the revamped photoshop workshop that was offered as part of the technology challenge, attendees brought family photos or film and learned how to edit and experiment with their photos and burn dvd copies. this time, the class was full: the same computer program that before drew only a few people was now exciting and useful. focusing on employees’ personal interests in learning new software, instead of just on teaching the software, better motivated staff and faculty to attend the training. motivation as stated by ota et al., adults are motivated by external factors but are usually more motivated by internal fac- tors: “adults are responsive to some external motivators (e.g., better job, higher salaries), but the most potent motivators are internal (e.g., desire for increased job satisfaction, self-esteem).” on the entrance survey, par- ticipants were given the opportunity to comment on their reasons for participating in the challenge. the gift card, an example of an external motivation, was frequently cited as an important motivation. but many also com- mented on more internal motivations: “it’s important to my job to stay proficient in new technologies and i’d like to stay current”; “i feel that i need to be up-to-date or meetings i could complete some of the challenges.” employees could also determine how much or how little of the challenge they wanted to complete: many reached well over the thousand-point goal, while others fell a little short. participants began at different skill levels, and thus could use the time and resources allotted to explore basic or more advanced topics according to their needs and interests. garrison had noted the importance of providing resources and feedback in self-directed learning. the techopoly website provided resources (such as specific blogs or websites to visit) and instructions on how to use and access technology within the library. hbll also hired a student to assist staff and faculty one-on-one by explain- ing answers to their questions about technology and teaching other skills he thought may be relevant to their initial problem. the entrance and exit surveys provided opportunities for self-reflection and self-evaluation by questioning the participants’ use of technology before the challenge and asking them to evaluate their proficiency in technology after the challenge. use of experience the use of experience as a source of learning is impor- tant to adult learners: “the richest resource for learning resides in adults themselves; therefore, tapping into their experiences through experiential techniques (discussions, simulations, problem-solving activities, or case methods) is beneficial.” the small-group discussions and one-on- one problem solving made available to hbll employees certainly fall into these categories. small-group classes are one of the best ways to encourage adults to share and validate their experiences, and doing so increases retention and application of new information. the trainings and challenges encouraged participants to make use of their work and personal experiences by connecting the topic to work or home application. for example, one session discussed how blogs relate to libraries, and another helped participants learn adobe photoshop skills by editing per- sonal photographs. need to know adult learners are more successful when they desire and recognize a need for new knowledge or skills. the role of a trainer is to help learners recognize this “need to know” by “mak[ing] a case for the value of learning.” hbll used the generational survey and presurvey to develop a need and desire to learn. the results of the generational survey, which demonstrated a gap in technology use between librarians and students, were presented and discussed at a meeting held before the initiation of the technology challenge to help staff and faculty under- stand why it was important to learn . technology. results of the presurvey showed that staff and faculty bridging the gap: self-directed staff technology training | quinney, smith, and galbraith statistical reports or working with colleagues from other libraries.” ■■ “i learned how to set up a server that i now maintain on a semi-regular basis. i learned a lot about sfx and have learned some perl programming language as well that i use in my job daily as i maintain sfx.” ■■ “the new oclc client was probably the most sig- nificant. i spent a couple of days in an online class learning to customize the client, and i use what i learned there every single day.” ■■ “i use google docs frequently for one of the projects i am now working on.” participants also indicated weaknesses in the technology challenge. almost percent of those who completed the challenge reported that it was too easy. this is a valid point—the challenge was designed to be easy so as not to intimidate staff or faculty who are less familiar with technology. it is important to note that these comments came from those who completed the challenge—other participants may have found the tasks and mini-challenges more difficult. the goal was to provide an introduction to web . , not to train experts. however, a greater range of tasks and challenges could be provided in the future to allow staff and faculty more self- direction in selecting goals relevant to their experience. to encourage staff and faculty to attend sponsored training sessions as part of the challenge, hbll leaders decided to double points for time spent at these classes. this certainly encouraged participation, but it lead to “point inflation”—perhaps being one reason why so many reported that the challenge was too easy to com- plete. the doubling of points may also have encouraged staff to spend more time in workshops and less time practicing or applying the skills learned. a possible solu- tion would be offering . points, or offering a set number of points for attendance instead of counting per minute. it also may have been informative for purpose of analy- sis to have surveyed both those who did not complete the challenge as well as those who chose not to participate. because the presurvey indicated that time was the biggest deterrent to learning and incorporating new technology, we assume that many of those who did not participate or who did not complete the challenge felt that they did not have enough time to do so. there is definitely potential for further investigation into why library staff would not want to participate in a technology training program, what would motivate them to participate, and how we could redesign the technology challenge to make it more appeal- ing to all of our staff and faculty. several library employees have requested that hbll sponsor another technology challenge program. because of the success of the first and because of continuing inter- est in technology training, we plan to do so in the future. we will make changes and adjustments according to the on technology in order to effectively help patrons”; “to identify and become comfortable with new technologies that will make my work more efficient, more presentable, and more accurate.” ■■ lifelong learning staff and faculty responded favorably to the training. none of the participants who took the exit survey disliked the challenge; percent even reported that they strongly liked it. ninety-five percent reported that they enjoyed the pro- cess of learning new technology, and percent reported that they were willing to participate in another technology challenge—thus suggesting success in the goal of encour- aging lifelong technology learning. the exit survey results indicate that after completing the challenge, staff and faculty are more motivated to continue learning—which is exactly what hbll leaders hoped to accomplish. eighty-nine percent of the partici- pants reported that their desire to learn new technology had increased, and percent reported that they are now able to learn new technology faster after completing the technology challenge. ninety-seven percent claimed that they were more likely to incorporate new technology into home or work use, and percent said they recognized the importance of staying on top of emerging technolo- gies. participants commented that the training increased their desire to learn. one observed, “i often need a chal- lenge to get motivated to do something new,” and another participant reported feeling “a little more comfortable trying new things out.” the exit survey asked participants to indicate how they now use technology. one employee keeps a blog for her daughter’s dance company, and another said, “i’m on my way to a full-blown googlereader addiction.” another participant applied these new skills at home: “i’m not so afraid of exploring the computer and other software programs. i even recently bought a computer for my own personal use at home.” the technology challenge was also successful in helping employees better serve patrons: “i can now better direct patrons to services that i would otherwise not have known about, such as streaming audio and video and e-book read- ers.” another participant felt better connected to student patrons: “i understand the students better and the things they use on a daily basis.” staff and faculty also found their new skills applicable to work beyond patron interaction, and many listed spe- cific examples of how they now use technology at work: ■■ “i have attended a few microsoft office classes that have helped me tremendously in doing my work more efficiently, whether it is for preparing monthly information technology and libraries | december . richard t. sweeny, “reinventing library buildings and services for the millennial generation,” library administration & management , no. ( ): . . win shish and martha allen, “working with generation- d: adopting and adapting to cultural learning and change,” library management , no. / ( ): . . sweeney, “reinventing library buildings,” . . shish and allen, “working with generation-d,” . . ibid., . . stephen abram, “social libraries: the librarian . pheonomenon,” library resources & technical services , no. ( ): . . ibid. . jack m. maness “library . theory: web . and its implications for libraries,” webology , no. ( ), http:// www.webology.ir/ /v n /a .html?q=link:webology.ir/ (accessed jan. , ). . ibid., under “blogs and wikis,” para. . . laurel ann clyde, “library weblogs,” library manage- ment , no. / ( ): – ; maness, “library . . theory.” . see matthew m. bejune, “wikis in libraries,” information technology & libraries , no. ( ): – ; darlene fichter, “the many forms of e-collaboration: blogs, wikis, portals, groupware, discussion boards, and instant messaging,” online: exploring technology & resources for information professionals , no. ( ): – ; maness, “library . theory.” . mary ellen bates, “can i facebook that?” online: explor- ing technology and resources for information professionals , no. ( ): ; sarah elizabeth miller and lauren a. jensen, “con- necting and communicating with students on facebook,” com- puters in libraries , no. ( ): – . . clyde, “library weblogs,” . . maness, “library . theory.” . fichter, “many forms of e-collaboration,” . . sweeney, “reinventing library buildings”; bates, “can i facebook that?” . bates, “can i facebook that?” . . abram, “social libraries,” . . ibid., . . ibid., . . shish and allen, “working with generation-d,” . . abram, “social libraries,” . . helene blowers and lori reed, “the c’s of our sea change: plans for training staff, from core competencies to learning . ,” computers in libraries , no. ( ): . . helene blowers, learning . , , http://plcmclearning .blogspot.com (accessed jan. , ). . for examples, see ilana kingsley and karen jensen, “learning . : a tool for staff training at the university of alaska fairbanks rasmuson,” the electronic journal of academic & special librarianship , no. ( ), http://southernlibrari- anship.icaap.org/content/v n /kingsley_i .html (accessed jan. , ); beverly simmons, “learning ( . ) to be a social library,” tennessee libraries , no. ( ): – . . for examples, see christine mackenzie, “creating our future: workforce planning for library . and beyond,” aus- tralasian public libraries & information services , no. ( ): – ; liisa sjoblom, “embracing technology: the deschutes public library’s learning . program,” ola quarterly , no. ( ): – ; hui-lan titango and gail l. mason, “learning library . : things @ scpl,” library management , no. / feedback we have received, and continue to evaluate it and improve it based on survey results. the purpose of a second technology challenge would be to reinforce what staff and faculty have already learned, to teach new skills, and to help participants remember the importance of life- long learning when it comes to technology. ■■ conclusion hbll’s self-directed technology challenge was success- ful in teaching technology skills and in promoting lifelong learning—as well as in fostering the development of librarian . . abram listed key characteristics and duties of librarian . , including learning the tools of web . ; connecting people, technology, and information; embrac- ing “nontextual information and the power of pictures, moving images, sight, and sound”; using the latest tools of communication; and understanding the “emerging roles and impacts of the blogosphere, web syndicasphere, and wikisphere.” survey results indicated that hbll employees are on their way to developing these attri- butes, and that they are better equipped with the skills and tools to keep learning. like plcmc’s learning . , the technology challenge could be replicated in libraries of various sizes. obviously an exact replication would not be feasible or appropriate for every library—but the basic ideas, such as the prin- ciples of andragogy and self-directed learning could be incorporated, as well as the daily time requirement or the use of surveys to determine weaknesses or interests in technology skills. whatever the case, there is a great need for library staff and faculty to learn emerging technolo- gies and to keep learning them as technology continues to change and advance. but the most important benefit of a self-directed train- ing program focusing on lifelong learning is effective employee development. the goal of any training pro- gram is to increase work productivity—and as employees become more productive and efficient, they are happier and more excited about their jobs. on the exit survey, one participant expressed initially feeling hesitant about the technology challenge and feared that it would increase an already hefty workload. however, once the challenge began, the participant enjoyed “taking the time to learn about new things. i feel i am a better person/librarian because of it.” and that, ultimately, is the goal—not only to create better librarians, but also to create better people. notes . robert h. mcdonald and chuck thomas, “disconnects between library culture and millennial generation values,” educause quarterly , no. ( ): . bridging the gap: self-directed staff technology training | quinney, smith, and galbraith ers,” journal of extension ( ), http://www.joe.org/ joe/ december/tt .php (accessed jan. , ); wayne g. west, “group learning in the workplace,” new directions for adult and continuing education ( ): – . . ota et al., “needs of learners.” . d. r. garrison, “self-directed learning: toward a com- prehensive model,” adult education quarterly ( ): . . ibid. . ota et al., “needs of learners,” under “needs of the adult learner,” para. . . ota et al., “needs of learners”; west, “group learning.” . ota et al., “needs of learners,” under “needs of the adult learner,” para. . . ibid., para. . . ibid., para . . abram, “social library,” – . ( ): – ; illinois library association, “continuous improve- ment: the transformation of staff development,” the illinois library association reporter , no. ( ): – ; and thomas simpson, “keeping up with technology: orange county library embraces . ,” florida libraries , no. ( ): – . . sharan b. merriam, “andragogy and self-directed learn- ing: pillars of adult learning theory,” new directions for adult & continuing education ( ): – . . malcolm shepherd knowles, the modern practice of adult education: from pedagogy to andragogy (new york: cambridge books, ). . jovita ross-gordon, “adult learners in the classroom,” new directions for student services ( ): – . . merriam, “pillars of adult learning”; ross-gordon, “adult learners.” . carrie ota et al., “training and the needs of learn- appendix a. technology challenge “mini challenges” technology challenge participants had the opportunity to complete fifteen of twenty mini-challenges to become eligible to win a second gift certificate to the campus bookstore. below are some examples of technology mini-challenges: . read a library or a technology blog . listen to a library podcast . check out a book from circulation’s new self-checkout machine . complete an online copyright tutorial . catalog some books on librarything . read an e-book with sony ebook reader or amazon kindle . scan photos or copy them from a digital camera and then burn them onto a cd . backup data . change computer settings . schedule meetings with microsoft outlook . create a page or comment on a page on the library’s intranet wiki . use one of the library’s music databases to listen to music . use wordpress or blogger to create a blog . post a photo on a blog . use google reader or bloglines to subscribe to a blog or news page using rss . reserve and check out a digital camera, camcorder, dvr, or slide scanner from the multimedia lab and create some- thing with it . convert media on the analog media racks . edit a family photograph using photo-editing software . attend a class in the multimedia lab . make a phone call using skype information technology and libraries | december how did you like the technology challenge overall? answer response percent strongly disliked disliked liked strongly liked how did you like the reporting system used for the technology challenge (the techopoly game)? answer response percent strongly disliked disliked liked strongly liked would you participate in another technology challenge? answer response percent yes no what percentage of time did you spend using the following methods of learning? (participants were asked to allocate points among the categories) category average response instructor-led large group . instructor-led small group one-on-one instruction . web tutorial . self-learning (books, articles) . dvds . small group discussion . large group discussion . other . i am more likely to incorporate new technology into my home or work life. answer response percent strongly disagree disagree agree strongly agree i enjoy the process of making new technology a part of my work or home life. answer response percent strongly disagree disagree agree strongly agree after completing the technology challenge, my desire to learn new technologies has increased. answer response percent strongly disagree disagree agree strongly agree i feel i now learn new technologies more quickly. answer response percent strongly disagree disagree agree strongly agree appendix b. exit survey results bridging the gap: self-directed staff technology training | quinney, smith, and galbraith open-ended questions ■■ what would you change about the technology chal- lenge? ■■ what did you like about the technology challenge? ■■ what technologies were you introduced to during the technology challenge that you now use on a regular basis? ■■ in what was do you feel the technology challenge has benefited you the most? how much more proficient do you feel in . . . category not any somewhat a lot hardware % % % software % % % internet resources % % % library technology % % % in order for you to succeed in your job, how important is keeping abreast of new technologies to you? answer response percent not important important very important information technology and libraries | december margaret brown-sica, jeffrey beall, and nina mchale next-generation library catalogs and the problem of slow response time and librarians will benefit from knowing what typical and acceptable response times are in online catalogs, and this information will assist in the design and evaluation of library discovery systems. this study also looks at bench- marks in response time and defines what is unacceptable and why. when advanced features and content in library catalogs increase response time to the extent that users become disaffected and use the catalog less, nextgen cata- logs represent a step backward, not forward. in august , the auraria library launched an instance of the worldcat local product from oclc, dubbed worldcat@auraria. the library’s traditional catalog—named skyline and running on the innovative interfaces platform—still runs concurrently with worldcat@auraria. because worldcat local currently lacks a library circulation module that the library was able to use, the legacy catalog is still required for its circulation functionality. in addition, skyline contains marc records from the serialssolution marc prod- uct. since many of these records are not yet available in the oclc worldcat database, these records are being maintained in the legacy catalog to enable access to the library’s extensive collection of online journals. almost immediately upon implementation of worldcat local, many library staff began to express concern about the product’s slow response time. they bemoaned its slowness both at the reference desk and during library instruction sessions. few of the discus- sions of the product’s slow response time evaluated this weakness in the context of its advanced features. several of the reference and instruction librarians even stated that they refused to use it any longer and that they were not recommending it to students and faculty. indeed, many stated that they would only use the legacy skyline catalog from then on. therefore we decided to analyze the product’s response time in relation to the legacy catalog. we also decided to further our study by examining response time in library catalogs in general, including several different online catalog products from different vendors. ■■ response time the term response time can mean different things in dif- ferent contexts. here we use it to mean the time it takes for all files that constitute a single webpage (in the case of testing performed, a permalink to a bibliographic record) to travel across the internet from a web server to the computer on which the page is to be displayed. we do not include the time it takes for the browser to render the page, only the time it takes for the files to arrive to the requesting computer. typically, a single webpage is made of multiple files; these are sent via the internet from a web response time as defined for this study is the time that it takes for all files that constitute a single webpage to travel across the internet from a web server to the end user’s browser. in this study, the authors tested response times on queries for identical items in five different library catalogs, one of them a next-generation (nextgen) catalog. the authors also discuss acceptable response time and how it may affect the discovery process. they suggest that librarians and vendors should develop standards for acceptable response time and use it in the product selec- tion and development processes. n ext-generation, or nextgen, library catalogs offer advanced features and functionality that facilitate library research and enable web . features such as tagging and the ability for end users to create lists and add book reviews. in addition, individual catalog records now typically contain much more data than they did in earlier generations of online catalogs. this additional data can include the previously mentioned tags, lists, and reviews, but a bibliographic record may also con- tain cover images, multiple icons and graphics, tables of contents, holdings data, links to similar items, and much more. this additional data is designed to assist catalog users in the selection, evaluation, and access of library materials. however, all of the additional data and features have the disadvantage of increasing the time it takes for the information to flow across the internet and reach the end user. moreover, the code that handles all this data is much more complex than the coding used in earlier, traditional library catalogs. slow response time has the potential to discourage both library patrons from using the catalog and library staff from using or recommending it. during a reference interview or library instruction ses- sion, a slow response time creates an awkward lull in the process, a delay that decreases confidence in the mind of library users, especially novices who are accustomed to the speed of an open internet search. the two-fold purpose of this study is to define the concept of response time as it relates to both traditional and nextgen library catalogs and to measure some typical response times in a selection of library catalogs. libraries margaret brown-sica (margaret.brown-sica@ucdenver.edu) is assistant professor, associate director of technology strat- egy and learning spaces, jeffrey beall (jeffrey.beall@ucden- ver.edu) is assistant professor, metadata librarian, and nina mchale (nina.mchale@ucdenver.edu) is assistant professor, web librarian, university of colorado denver. next-generation library catalogs | brown-sica, beall, and mchale mathews posted an article called “ next gen library catalogs and students: their initial impressions.” here he shares student impressions of several nextgen catalogs. regarding slow response time mathews notes, “lots of comments on slowness. one student said it took more than ten seconds to provide results. some other comments were: ‘that’s unacceptable’ and ‘slow-motion search, typical library.’” nagy and garrison, on lauren’s library blog, emphasized that any “cross-silo federated search” is “as slow as the slower silos.” any search inter- face is as slow as the slowest database from which it pulls information; however, that does not make users more likely to wait for search results. in fact, many users will not even know—or care—what is happening behind the scenes in a nextgen catalog. the assertion that slow response time makes well- intentioned improvements to an interface irrelevant is supported by an article that analyzes the development of semantic web browsers. frachtenberg notes that users, however, have grown to expect web search engines to provide near-instantaneous results, and a slow search engine could be deemed unusable even if it provides highly relevant results. it is therefore imperative for any search engine to meet its users’ interactivity expectations, or risk losing them. this is not just a library issue. users expect a fast response to all web queries, and we can learn from studies on general web response time and how it affects the user experience. huang and fong-ling help explain different user standards when using websites. their research suggests that “hygiene factors” such as “navigation, information display, ease of learning and response time” are more important to people using “utilitarian” sites to accomplish tasks rather than “hedo- nistic” sites. in other words, response time importance increases when the user is trying to perform a task— such as research—and possibly even more for a task that may be time sensitive—such as trying to complete an assignment for class. ■■ method for testing response time in an assortment of library cat- alogs, we used the websitepulse service (http://www .websitepulse.com). websitepulse provides in-depth website and server diagnostic services that are intended to save e-business customers time and money by reporting errors and web server and website performance issues to clients. a thirty-day free trial is available for potential cus- tomers to review the full array of their services; however, the free web page test, available at http://www.website server and arrive sequentially at the computer where the request was initiated. while the world wide web consortium (w c) does not set forth any particular guidelines regarding response time, go-to usability expert jakob nielsen states that “ . second is about the limit for having the user feel that the system is reacting instantaneously.” he further posits that . second is “about the limit for the user’s flow of thought to stay uninterrupted, even though the user will notice the delay.” finally, he asserts that: seconds is about the limit for keeping the user’s attention focused on the dialogue. for longer delays, users will want to perform other tasks while waiting for the computer to finish, so they should be given feedback indicating when the computer expects to be done. feedback during the delay is especially impor- tant if the response time is likely to be highly variable, since users will then not know what to expect. even though this advice dates to , nielsen noted even then that it had “been about the same for many years.” ■■ previous studies the chief benefit of studying response time is to estab- lish it as a criterion for evaluating online products that libraries license and purchase, including nextgen online catalogs. establishing response-time benchmarks will aid in the evaluation of these products and will help libraries convey the message to product vendors that fast response time is a valuable product feature. long response times will indicate that a product is deficient and suffers from poor usability. it is important to note, however, that sometimes library technology environments can be at fault in lengthening response time as well; in “playing tag in the dark: diagnosing slowness in library response time,” brown-sica diag- nosed delays in response time by testing such variables as vendor and proxy issues, hardware, bandwidth, and network traffic. in that case, inadequate server specifi- cations and settings were at fault. while there are many articles on nextgen catalogs, few of them discuss the issue of response time in rela- tion to their success. search slowness has been reported in library literature about nextgen catalogs’ metasearch cousins, federated search products. in a review of federated search tools metalib and webfeat, chen noted that “a federated search could be dozens of times slower than google.” more comments about the negative effects of slow response time in nextgen catalogs can be found in popular library technology blogs. on his blog, information technology and libraries | december ■■ findings: skyline versus worldcat@auraria in figure , the bar graph shows a sample load time for the permalink to the bibliographic record for the title hard lessons: the iraq reconstruction experience in skyline, auraria’s traditional catalog load time for the page is pulse.com/corporate/alltools.php, met our needs. to use the webpage test, simply select “web page test” from the dropdown menu, input a url—in the case of the testing done for this study, the perma- link for one of three books (see, for example, figure )—enter the validation code, and click “test it.” websitepulse returns a bar graph (figure ) and a table (figure ) of the file activity from the server sending the composite files to the end user ’s web browser. each line represents one of the files that make up the rendered webpage. they load sequentially, and the bar graph shows both the time it took for each file to load and the order in which the files were received. longer seg- ments of the bar graph provide visual indication of where a slow-loading webpage might encounter sticking points—for example, wait- ing for a large image file or third-party content to load. accompanying the bar graph is a table describing the file transmissions in more detail, including dns, connection, file redirects (if applicable), first and last bytes, file trans- mission times, and file sizes. figure . permalink screen shot for the record for the title hard lessons in auraria library’s skyline catalog figure . websitepulse webpage test bar graph results for skyline (traditional) catalog record figure . websitepulse webpage test table results for skyline (traditional) catalog record next-generation library catalogs | brown-sica, beall, and mchale requested at items , , , , , and . the third parties include yahoo! api services, the google api ser- vice, recaptcha, and addthis. recaptcha is used to provide security within worldcat local with opti- cal character recognition images (“captchas”), and the addthis api is used to provide bookmarking function- ality. at number , a connection is made to the auraria library web server to retrieve a logo image hosted on the web server. at number , the cover photo for hard lessons is retrieved from an oclc server. the files listed in figure details the complex process of web brows- ers’ assembly of them. each connection to third-party content, while all relatively short, allows for addi- tional features and functionality, but lengthens overall response. as figure shows, the response time is slightly more than seconds, which, according to nielsen, “is about the limit for keeping the user ’s attention focused on the dialogue.” while widgets, third-party content, and other web . tools add desirable content and functionality to the library’s catalog, they also do slow response time considerably. the total file size for the bibliographic record in worldcat@auraria—compared to skyline’s . kb—is . kb. as will be shown in the test results below for the catalog and nextgen catalog products, bells and whistles added to traditional . seconds total. the record is composed of a total of fourteen items, including image files (gifs), cascad- ing style sheet (css) files, and javascript (js) files. as the graph is read downward, the longer segments of the bars reveal the sticking points. in the case of skyline, the nine image files, two css files, and one js file loaded quickly; the only cause for concern is the red line at item four. this revealed that we were not taking advantage of the option to add a favicon to our iii catalog. the web librarian provided the ils server technician with the same favi- con image used for the library’s website, correcting this issue. the skyline catalog, judging by this data, falls into nielsen’s second range of user expectations regarding response time, which is more than one second, or “about the limit for the user’s flow of thought to stay uninter- rupted, even though the user will notice the delay.” further detail is provided in figure ; this table lists each of the webpage’s component files, and various times asso- ciated with the delivery of each file. the column on the right lists the size in kilobytes of each file. the total size of the combined files is . kb. in contrast to skyline’s meager files, worldcat local requires items to assemble the webpage (figure ) for the same bibliographic record. figures and show that this includes css files, javascript files, and images files (gifs and pngs). no item in particular slows down the overall process very much; the longest- loading item is number , which is a wait for third-party content, a connection to yahoo!’s user interface (yui) api service. additional third-party content is being figure . permalink screen shot for the record for the title hard lessons in worldcat@auraria figure . websitepulse webpage test bar graph results for worldcat@auraria record information technology and libraries | december total time for each permalinked bibliographic record to load as reported by the websitepulse tests; this number appears near the lower right-hand corner of the tables in figures , , , , and . we selected three books that were each held by all five of our test sites, verifying that we were search- ing the same three bibliographic records in each of the online catalogs by looking at the oclc number in the records. each of the catalogs we tested has a permalink feature; this is a stable url that always points to the same record in each catalog. using a permalink approximates conducting a known-item search for that item from a catalog search screen. we saved these links and used them in our searches. the bib- liographic records we tested were for these books; the permalinks used for testing follow the books: book : hard lessons: the iraq reconstruction experience. washington, d.c.: special inspector general, iraq reconstruction, (oclc number ). permalinks used: ■■ worldcat@auraria: http://aurarialibrary.worldcat .org/oclc/ ■■ skyline: http://skyline.cudenver.edu/record=b ~s ■■ lcoc: http://lccn.loc.gov/ ■■ ut austin: http://catalog.lib.utexas.edu/record= b ~s ■■ usc: http://library.usc.edu/uhtbin/cgisirsi/ x/ / / ?searchdata = {ckey} book : ehrenreich, barbara. nickel and dimed: on (not) getting by in america. st ed. new york: metropolitan, (oclc number ). permalinks used: ■■ worldcat@auraria: http://aurarialibrary.worldcat .org/oclc/ ■■ skyline: http://skyline.cudenver.edu/record=b ~s ■■ lcoc: http://lccn.loc.gov/ ■■ ut austin: http://catalog.lib.utexas.edu/record= b ~s ■■ usc: http://library.usc.edu/uhtbin/cgisirsi/ x/ / / ?searchdata = {ckey} book : langley, lester d. simón bolívar: venezuelan rebel, american revolutionary. lanham: rowman & littlefield catalogs slowed response time considerably, even dou- bling it in one case. are they worth it? the response of auraria’s reference and instruction staff seems to indi- cate that they are not. ■■ gathering more data: selecting the books and catalogs to study to broaden our comparison and to increase our data collection, we also tested three other non-auraria cata- logs. we designed our study to incorporate a number of variables. we decided to link to bibliographic records for three different books in the five different online catalogs tested. these included skyline and worldcat@auraria as well three additional online public access catalog products, for a total of two instances of innovative interfaces products, one of a voyager catalog, and one of a sirsidynix catalog. we also selected online catalogs in different parts of the country: worldcatlocal in ohio; skyline in denver; the library of congress’ online catalog (lcoc) in washington, d.c.; the university of texas at austin’s (ut austin) online catalog; and the university of southern california’s (usc) online catalog, named homer, in los angeles. we also did our testing at different times of the day. one book was tested in the morning, one at midday, and one in the afternoon. websitepulse performs its webpage tests from three different locations in seattle, munich, and brisbane; we selected seattle for all of our tests. we recorded the figure . websitepulse webpage test table results for worldcat@auraria record next-generation library catalogs | brown-sica, beall, and mchale .org/oclc/ ■■ skyline: http://skyline.cudenver.edu/record=b ~s ■■ lcoc: http://lccn.loc.gov/ ■■ ut austin: http://catalog.lib.utexas.edu/record= b ~s ■■ usc: http://library.usc.edu/uhtbin/cgisirsi/ x/ / / ?searchdata = {ckey} we gathered the data for thirteen days in early november , an active period in the middle of the semester. for each test, we recorded the response time total in seconds. the data is displayed in tables – . we searched bibliographic records for three books in five library catalogs over thirteen days ( x x ) for a total of response time measurements. the websitepulse data is calculated to the ten thousandth of a second, and we recorded the data exactly as it was presented. publishers, c (oclc number ). permalinks used: ■■ worldcat@auraria: http://aurarialibrary.worldcat table . response times for book response time in seconds day wor ld- cat skyline lc ut austin usc . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . average . . . . . table . response times for book response time in seconds day world- cat skyline lc ut austin usc . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . average . . . . . table . response times for book response time in seconds day world- cat skyline lc ut austin usc . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . average . . . . . table . averages response time in seconds book world- cat skyline lc ut austin usc book . . . . . book . . . . . book . . . . . average . . . . . information technology and libraries | december university of colorado denver: skyline (innovative interfaces) as previously mentioned, the traditional catalog at auraria library runs on an innovative interfaces integrated library system (ils). testing revealed a missing favicon image file that the web server tries to send each time (item in figure ). however, this did not negatively affect the response time. the catalog’s response time was good, with an aver- age of . seconds, giving it the fastest average time among all the test sites in the testing period. as figure shows, however, skyline is a typical legacy catalog that is designed for a traditional library environment. library of congress: online catalog (voyager) the average response time for the lcoc was . ■■ results the data shows the response times for each of the three books in each of the five online catalogs over the thirteen- day testing period. the raw data was used to calculate averages for each book in each of the five online catalogs, and then we calculated averages for each of the five online catalogs (table ). the averages show that during the testing period, the response time varied between . seconds for the skyline library catalog in denver to . seconds for worldcat@auraria, which has its servers in ohio. university of colorado denver: worldcat@auraria worldcat@auraria was routinely over nielsen’s ten- second limit, sometimes taking as long as twenty sec- onds to load all the files to generate a single webpage. as previously discussed, this is due to the high number and variety of files that make up a single bibliographic record. the files sent also include cover images, but they are small and do not add much to the total time. after our tests on worldcat@auraria were conducted, the site removed one of the features on pages for individual resources, namely the “similar items” feature. this fea- ture was one of the most file-intensive on a typical page, and its removal should speed up page loads. however, worldcat@auraria had the highest average response time by far of the five catalogs tested. figure . permalink screen shot for the record for the title hard lessons in the library of congress online catalog figure . websitepulse webpage test bar graph results for library of congress online catalog record figure . websitepulse webpage test table results for library of congress online catalog record next-generation library catalogs | brown-sica, beall, and mchale item is a script, that while hosted on the ils server, que- ries amazon.com to return cover image art (figures – ). the average response time for ut austin’s catalog was . seconds. this example demonstrates that response times for traditional (i.e., not nextgen) catalogs can be slowed down by additional content as well. university of southern california: homer (sirsidynix) the average response time for usc’s homer catalog was . seconds, making it the second slowest after seconds. this was the second fastest average among the five catalogs tested. while, like skyline, the bibliographic record page is sparsely decorated (figure ), this pays dividends in response time, as there are only two css files and three gif files to load after the html content loads (figure ). figure shows that initial connection time is the longest factor in load time; however, it is still short enough to not have a negative effect. total file size is . kb. as with skyline, the page itself (figure ) is not particularly end-user friendly to nonlibrarians. university of texas at austin: library catalog (innovative interfaces) ut austin, like auraria library, runs an innovative interfaces ils. the library catalog also includes book cover images, one of the most attractive nextgen features (figure ), and as shown in figure , third-party content is used to add features and functionality (items and ). ut austin’s catalog uses a google javascript api (item in figure ) and librarything’s catalog enhancement prod- uct, which can add book recommendations, tag browsing, and alternate editions and translations. total content size for the bibliographic record is considerably larger than skyline and the lcoc at . kb. it appears as though inclusion of cover art nearly doubles the response time; figure . permalink screen shot for the record for the title hard lessons in university of texas at austin’s library catalog figure . websitepulse webpage test bar graph results for university of texas at austin’s library catalog record figure . websitepulse webpage test table results for university of texas at austin’s library catalog record information technology and libraries | december completed. added functionality and features in library search tools are valuable, but there is a tipping point when these features slow down a product’s response time to where users find the search tool too slow or unreliable. based on the findings of this study, we recom- mend that libraries adopt web response time standards, such as those set forth by nielsen, for evaluating vendor search products and creating in-house search products. commercial tools like websitepulse make this type of data collection simple and easy. testing should be con- ducted for an extended period of time, preferably during a peak period—i.e., during a busy time of the semes- ter for academic libraries. we further recommend that reviews of electronic resources add response time as an worldcat@auraria, and the slowest among the tradi- tional catalogs. this sirsidynix catalog appears to take a longer time than the other brands of catalogs to make the initial connection to the ils; this accounts for much of the slowness (see figures and ). once the initial connection is made, however, the remaining content loads very quickly, with one exception: item (see fig- ure ), which is a connection to the third-party provider syndetic solutions, which provides cover art, a summary, an author biography, and a table of contents. while the display of this content is attractive and well-integrated to the catalog (figure ), it adds . seconds to the total response time. also, as shown in item and , usc’s homer uses the addthis service to add bookmarking enhancements to the catalog. total combined file size is . kb, with the bulk of the file size ( kb) coming from the initial connection (item in figure ). ■■ conclusion an eye-catching interface and valuable content are lost on the end user if he or she moves on before a search is figure . permalink screen shot for the record for the title hard lessons in homer, the university of southern california’s catalog figure . websitepulse webpage test bar graph results for homer, the university of southern california’s catalog figure . websitepulse webpage test table results for homer, the university of southern california’s catalog next-generation library catalogs | brown-sica, beall, and mchale . ibid. . margaret brown-sica. “playing tag in the dark: diagnos- ing slowness in library response time,” information technology & libraries , no. ( ): – . . xiaotian chen, “metalib, webfeat, and google: the strengths and weaknesses of federated search engines com- pared with google,” online information review , no. ( ): . . brian mathews, “ next gen library catalogs and stu- dents: their initial impressions,” online posting, may , , the ubiquitous librarian blog, http://theubiquitouslibrarian .typepad.com/the_ubiquitous_librarian/ / / -next-gen- library-catalogs-and- -students-their-initial-impressions.html (accessed feb. , ) . andrew nagy and scott garrison, “next-gen catalogs are only part of the solution,” online posting. oct. , , lauren’s library blog, http://laurenpressley.com/library/ / /next -gen-catalogs-are-only-part-of-the-solution/ (accessed feb. , ). . eitan frachtenberg, “reducing query latencies in web search using fine-grained parallelism,” world wide web , no. ( ): – . . travis k huang and fu fong-ling, “understanding user interface needs of e-commerce web sites,” behaviour & information technology , no. ( ): – , http://www .informaworld.com/ . / (accessed feb. , ). . nielsen, usability engineering, . . ibid. evaluation criterion. additional research about response time as defined in this study might look at other search tools, to include article databases, and especially other metasearch products that collect and aggregate search results from several remote sources. further studies with more of a technological focus could include discussions of optimizing data delivery methods—again, in the case of metasearch tools from multiple remote sources—to reduce response time. finally, product designers should pay close attention to response time when designing information retrieval products that libraries purchase. ■■ acknowledgments the authors wish to thank shelley wendt, library data analyst, for her assistance in preparing the test data. references . jakob nielsen, usability engineering (san francisco: morgan kaufmann, ): . . ibid. . ibid. information technology and libraries | september lita committees and interest groups are being asked to step up to the table and develop action plans to imple- ment the strategies the lita membership have identified as crucial to the association’s ongoing success. members of the board are liaisons to each of the committees, and there is a board liaison to the interest groups. these indi- viduals will work with committee chairs, interest group chairs, and the membership to implement lita’s plan for the future. the committee and interest group chairs are being asked to contribute those actions plans by the ala midwinter meeting. they will be compiled and made available to all lita and ala members for their use through the lita website (http://lita.org) and ala connect (http://connect.ala.org). what is in it for you? lita is known for its leadership opportunities, continuing education, training, publica- tions, expertise in standards and information policy, and knowledge and understanding of current and cutting- edge technologies. lita provides you with opportunities to develop those leadership skills that you can use in your job and lifelong career. the skills working within a group of individuals to implement a program, influence standards and policy, collaborate with other ala divi- sions, and publish can be taken home to your library. your participation documents your value as an employee and your commitment to lifelong learning. in today’s work environment, employers look for staff with proven skills who have contributed to the good of the organization and the profession. lita needs your participation in developing and implementing continuing education programs, publish- ing articles and books, and illustrating by your actions why others want to join the association. how can you do that? volunteer for a committee, help develop a continu- ing education program, write an article, write a book, role model for others with your lita participation, and recruit. what does your association gain? a solid struc- ture to support its members in accomplishing the mission, vision, and strategic plan they identified as core for years to come. look for opportunities to participate and develop those skills. we will be working with committee and interest group chairs to develop meeting management tool kits over the next year, create opportunities to par- ticipate virtually, identify emerging leaders of all types, collaborate with other divisions, and provide input on national information policy and standards through ala’s office for information technology policy and other simi- lar organizations. if you want to be involved, be sure to let lita committee and interest group chairs, the board, and your elected officers know. c loud computing. web . or the semantic web. google editions. books in copyright and books out of copyright. born digital. digitized material. the reduction of stanford university’s engineering library book collection by percent. the publishing paradigm most of us know, and have taken for granted, has shifted. online databases came and we managed them. then cd-roms showed up and mostly went away. and, along came the internet, which we helped implement, use, and now depend on. how we deal with the current shifts happening in information and technology during the next five to ten years will say a great deal about how the library and information community reinvents itself for its role in the twenty-first century. this shift is different, and it will create both opportunities and challenges for everyone, including those who manage information and those who use it. as a reflection of the shifts in the information arena, lita is facing its own challenges as an association. it has had a long and productive role in the american library association (ala) dating back to . the talent among the association members is amazing, solid, and a tribute to the individuals who belong to and participate in lita. lita’s members are leaders to the core and recognized as standouts within ala as they push the edge of what information management means, and can mean. for the past three years, lita members, the board, and the executive committee have been working on a strategic plan for lita. that process has been described in michelle frisque’s “president’s message” (ital v. , no. ) and elsewhere. the plan was approved at the ala annual conference in washington, d.c. a plan is not cast in concrete. it is a dynamic, living document that provides the fabric that drives the association. why is this process important now more than ever? we are all dealing with the current recession. libraries are retrenching. people face challenges participating in the library field on various levels. the big information players on the national and international level are chang- ing the playing field. as membership, each of us has an opportunity to affect the future of information and tech- nology locally, nationally, and internationally. this plan is intended to ensure lita’s role as a “go to” place for people in the library, information, and technology fields well into the twenty-first century. karen j. starr (kstarr@nevadaculture.org) is lita president – and assistant administrator for library and develop- ment services, nevada state library and archives, carson city. karen j. starrpresident’s message: moving forward editorial board thoughts | dehmlow i n the age of the internet, google, and the nearly crushing proliferation of metadata, libraries have been struggling with how to maintain their relevance and survive in the face of shrinking budgets and misinformed questions about whether libraries still provide value. in case there was ever any question, the answer is “of course we do.” still, an evolving environment and changing con- text has motivated us to rethink what we do and how we do it. our response to the shifting environment has been to envision how libraries can provide the best value to our patrons despite an information ecosystem that duplicates (and to some extent replaces) services that have been a core part of our profession for ages. at the same time, we still have to deal with procedures for managing resources we acquire and license, and many of the systems and pro- cesses that have served us so well for so many years are not suitable for today’s environment. many have talked about the need to invest in the distinctive services we provide and unique collections we have (e.g., preserving the world’s knowledge and digi- tizing our unique holdings) as a means to add value to libraries. there are many other ways libraries create value for our users, and one of the best is for us to respond to needs that are specific to our organizations and users— specialized services, focused collections, contextualized discovery, all integrated into environments in which our patrons work, such as course management systems, google, etc. the library market has responded to many of our needs with ermss and next-generation resource management and discovery solutions. all of this is a good start, but like any solution that is designed to work for the greatest common denominator, they often leave a “desired functionality gap” because no one system can do everything for everyone, no development today can address all of the needs of tomorrow, and very rarely do all of the disparate systems integrate with each other. so where does that leave libraries? well, every prob- lem is an opportunity, and there are two important areas that libraries can invest in to ensure that they progress at the same pace as technology, their users, and the mar- ket: open systems that have application programmer interfaces (apis), and programmers. apis are a means to access the data and functionality of our vended or open- source systems using a program as opposed to the default interface. apis often take the shape of xml travelling in the same way that webpages do, accessed via a url, but they also can be as complex as writing code in the same language as the base system, for example software devel- opment kits (sdks). the key here is that apis provide a way to work with the data in our systems, be they back- end inventory or front-end discovery interfaces, in ways that weren’t conceived by the software developers. this flexibility enables organizations to respond more rapidly to changing needs. no matter which side of the open- source/vended solution fence you sit on, openness needs to be a fundamental part of any decision process for any new system (or information service) to avoid being stifled by vendor or open-source developer priorities that don’t necessarily reflect your own. the second opportunity is perhaps the more diffi- cult one given the state of library budgets and that the resources that are needed to hire programmers are higher than most other library staff. but having local program- ming skills easily accessible will be vital to our ability to address our users’ specific needs and change our internal processes as we need to. i think it is good to have at least one technical person who comes from an industry outside of libraries. they bring knowledge that we don’t neces- sarily have and fresh perspectives on how we do things. if it is not possible to hire a programmer, i would encourage technology managers to look closely at their existing staff, locate those in the organization who are able to think outside of the box, and provide some time and space for them to grow their skill set. i am not so obtuse as to suggest that anyone can be programmer—like any skill it requires a general aptitude and a fundamental inter- est—but i am a self-taught developer who had a technical aptitude and an strong desire to learn new things, and i suspect that there are many underutilized staff in librar- ies that with a little encouragement, mentoring, and some new technical knowledge, could easily work with apis and sdks, thereby opening the door for organizations to be nimble and responsive to both internal and external needs. i recognize that with heavy demands it can be difficult to give up some of these highly valued people’s time, but the payoff is overwhelmingly worth it. these days i can only chuckle at the doomsday predictions about libraries and the death of our services— google’s dominance in the search arena has never really made me worried that libraries would become irrelevant. we have too much that google does not, specifically licensed content that our users desire, and we have rela- tionships with our users that google will be incapable of having. i have confidence that what we have to offer will be valuable to our users for some time to come. however, it will take a willingness to evolve with our environment and to invest in skill sets that come at a premium even when it is difficult to do so. mark dehmlow editorial board thoughts: adding value in the internet age— libraries, openness, and programmers mark dehmlow (mdehmlow@nd.edu) is digital initiatives librar- ian, university of notre dame, notre dame, indiana. information technology and libraries | september development of such a mediation mechanism calls for an empirical assessment of various issues surrounding metadata-creation practices. the critical issues concerning metadata practices across distributed digital collections have been rela- tively unexplored. while examining learning objects and e-prints communities of practice, barton, currier, and hey point out the lack of formal investigation of the metadata- creation process. as will be discussed in the following section, some researchers have begun to assess the current state of descriptive practices, metadata schemata, and content standards. however, the literature has not yet developed to a point where it affords a comprehensive picture. given the propagation of metadata projects, it is important to continue to track changes in metadata-cre- ation practices while they are still in constant flux. such efforts are essential for adding new perspectives to digital library research and practices in an environment where metadata best practices are being actively sought after to aid in the creation and management of high-quality digital collections. this study examines the prevailing current state of metadata-creation practices in digital repositories, col- lections, and libraries, which may include both digitized and born-digital resources. using nationwide survey data, mostly drawn from the community of catalog- ing and metadata professionals, we seek to investigate issues in creating descriptive metadata elements, using controlled vocabularies for subject access, and propa- gating metadata and metadata guidelines beyond local environments. we will address the following research questions: . which metadata schema(ta) and content standard(s) are employed in individual digital repositories and collections? . which controlled vocabulary schema(ta) are used to facilitate subject access? . what criteria are applied in selecting metadata and controlled-vocabulary schema(ta)? . to what extent are mechanisms for exposing and sharing metadata integrated into current metadata- creation practices? in this article, we first review recent studies relating to current metadata-creation practices across digital col- lections. then we present the survey method employed to conduct this study, the general characteristics of survey participants, and the validity of the collected data, fol- lowed by the study results. we report on how metadata and controlled vocabulary schema(ta) are being used across institutions, and we present a data analysis of current metadata-creation practices. the final section summarizes the study and presents some suggestions for future studies. this study explores the current state of metadata-creation practices across digital repositories and collections by using data collected from a nationwide survey of mostly cataloging and metadata professionals. results show that marc, aacr , and lcsh are the most widely used metadata schema, content standard, and subject- controlled vocabulary, respectively. dublin core (dc) is the second most widely used metadata schema, followed by ead, mods, vra, and tei. qualified dc’s wider use vis-à-vis unqualified dc ( . percent versus . percent) is noteworthy. the leading criteria in selecting metadata and controlled-vocabulary schemata are collec- tion-specific considerations, such as the types of resources, nature of the collection, and needs of primary users and communities. existing technological infrastructure and staff expertise also are significant factors contributing to the current use of metadata schemata and controlled vocabularies for subject access across distributed digital repositories and collections. metadata interoperability remains a major challenge. there is a lack of exposure of locally created metadata and metadata guidelines beyond the local environments. homegrown locally added meta- data elements may also hinder metadata interoperability across digital repositories and collections when there is a lack of sharable mechanisms for locally defined extensions and variants. m etadata is an essential building block in facili- tating effective resource discovery, access, and sharing across ever-growing distributed digital collections. quality metadata is becoming critical in a networked world in which metadata interoperability is among the top challenges faced by digital libraries. however, there is no common data model that catalog- ing and metadata professionals can readily reference as a mediation mechanism during the processes of descriptive metadata creation and controlled vocabu- lary schemata application for subject description. the jung-ran park (jung-ran.park@ischool.drexel.edu) is assistant professor, college of information science and technology, drex- el university, philadelphia, and yuji tosaka (tosaka@tcnj.edu) is cataloging/metadata librarian, tcnj library, the college of new jersey, ewing, new jersey. jung-ran park and yuji tosaka metadata creation practices in digital repositories and collections: schemata, selection criteria, and interoperability metadata creation practices in digital repositories and collections | park and tosaka possible increase in the use of locally developed schemata as many projects added new types of nontextual digital objects that could not be adequately described by existing metadata schemata. there is a lack of research concerning the current use of content standards; however, it is reasonable to suspect that content-standards use exhibits patterns similar to that of metadata because of their often close association with particular metadata schemata. the oclc rlg sur- vey reveals that anglo-american cataloguing rules, nd edition (aacr )—the traditional cataloging rule that has most often been used in conjunction with marc—is the most widely used content standard ( percent). aacr is followed by describing archives: a content standard (dacs) with percent; descriptive cataloging of rare materials with percent; archives, personal papers, manuscripts (appm) with percent; and cataloging cultural objects (cco) with percent. in the same way as metadata schemata, there appears to be a concentration of a few controlled vocabulary schemata at research institutions. ma’s arl survey, for example, shows that the library of congress subject headings (lcsh) and name authority file (naf) were used by most survey respondents ( percent and percent, respectively). these two predominantly adopted vocabularies are followed by several domain-specific vocabularies, such as art and architecture thesaurus (aat), library of congress thesaurus for graphical materials (tgm) i and ii, getty thesaurus of geographic names (tgn), and the getty union list of artists names (ulan), which were used by between percent to more than percent of respondents. the oclc rlg survey reports similar results; however, nearly half of the oclc rlg survey respondents (n = ) indicated that they had also built and maintained one or more locally developed thesauri. while creating and sharing information about local metadata implementations is an important step toward increased interoperability, recent studies tend to paint a grim picture of current local documentation practices and open accessibility. in a nationwide study of institutional repositories in u.s. academic libraries, markey et al. found that only . percent of the survey participants with operational institutional repositories had imple- mented policies for metadata schemata and authorized metadata creators. the oclc rlg survey also high- lights limited collaboration and sharing of the metadata guidelines both within and across the institutions. it finds that even when there are multiple units creating metadata within the same institution, metadata-creation guidelines often are unlikely to be shared ( percent do not share; percent sometimes share). a mixed result is reported on the exposure of meta- data to outside service providers. in an arl survey, the university of houston libraries institutional repository ■■ literature review as evinced by the principles and practices of bib- liographic control through shared cataloging, successful resource access and sharing in the networked envi- ronment demands semantic interoperability based on accurate, complete, and consistent resource description. the recent survey by ma finds that the open archives initiative protocol for metadata harvesting (oai-pmh) and metadata crosswalks have been adopted by percent and percent of respondents, respectively. even though the sample comes only from sixty-eight association of research libraries (arl) member librar- ies, and the figures thus may be skewed higher than those of the entire population of academic libraries, there is little doubt that interoperability is a critical issue given the rapid proliferation of metadata schemata throughout digital libraries. while there is a variety of metadata schemata cur- rently in use for organizing digital collections, only a few of them are widely used in digital repositories. in her arl survey, ma reports that the marc format is the most widely used metadata schema ( percent), followed by encoded archival description (ead) ( percent), unqualified dublin core (dc) ( percent), and qualified dc ( percent). similarly, a member survey by oclc research libraries group (rlg) programs gath- ered information from eighteen major research libraries and cultural heritage institutions and also found that marc is the most widely used scheme ( percent), fol- lowed by ead ( percent), unqualified dc ( percent), and qualified dc ( percent). the different levels of use reported by these studies are probably due to different sample sizes and compositions, but results nonetheless suggest that metadata use at research institutions tends to rely on a small number of major schemata. there may in fact be much greater diversity in meta- data use patterns when the scope is expanded to include both research and nonresearch institutions. palmer, zavalina, and mustafoff, for example, tracked trends from through in metadata selection and application practices at more than digital collections developed through institute of museum and library services grants. they found that despite perceived limitations, use of dc is the most widespread, with more than half of the digital collections using it alone or in combination with other schemata. marc ranks second, with nearly percent using it alone or in combination. the authors found that the choice of metadata schema is largely influenced by practices at peer institutions and compatibility with a content management system. what is most striking, how- ever, is the finding that locally developed schemata are used as often as marc. there is a decline in the percent- age of digital projects using multiple metadata schemata (from percent to percent). yet the authors also saw a information technology and libraries | september ■■ method the objective of the research reported in this paper is to examine the current state of metadata-creation practices in terms of the creation of descriptive metadata elements, the use of controlled vocabularies for subject access, and the exposure of metadata and metadata guidelines beyond local environments. we conducted a web survey using websurveyor (now vovici: http://www.vovici .com). the survey included both structured and open- ended questions. it was extensively reviewed by members of an advisory board—a group of three experts in the field—and it was pilot-tested prior to being officially launched. the survey included many multiple-response questions that called for respondents to check all appli- cable answers. we recruited participants through survey invitation messages and subsequent reminders to the electronic mailing lists of communities of metadata and cataloging professionals. table shows the mailing lists employed for the study. we also sent out individual invitations and distributed flyers to selected metadata and cataloging ses- sions during the ala midwinter meeting, held that year in philadelphia. the survey attracted a large number of initial par- ticipants (n = , ), but during the sixty-two days from august to october , , we only received com- pleted responses via the survey management system. we suspect that the high incompletion rate ( . percent) stems from the fact that the subject matter may have been outside the scope of many participants’ job responsibili- ties. the length of the survey may also have been a factor in the incompletion rate. the profiles of respondents’ job titles (see table ) task force found that exposing metadata to oai-pmh service providers is an established practice used by nearly percent of the respondents. ma’s arl survey also reports the wide adoption of oai-pmh ( per- cent). these results underscore the virtual consensus on the critical importance of exposing metadata to achieve interoperability and make locally created metadata useful across distributed digital repositories and collections. by contrast, the oclc rlg survey shows that only one-tenth of the respondents stated that all non-marc metadata is exposed to oai harvesters, while percent indicated that only some of it was available. the prominent theme revealed by the oclc rlg survey is an “inward focus” in current metadata practices, marked by the “use of local tools to reach a generally local audience.” in summary, recent studies show that the current practice of metadata creation is problematic due to the lack of a mechanism for integrating various types of metadata schemata, content standards, and controlled vocabularies in ways that promote an optimal level of interoperability across digital collections and repositories. the problems are exacerbated in an environment where many institutions lack local documentation delineating the metadata-creation process. at the same time, researchers have only recently begun studying these issues, and the body of literature is at an incipient stage. the research that was done often targeted different populations, and sample sizes were different (some very small). in some cases the literature exhibits contradictory findings about issues surrounding metadata practices, increasing the difficulty in understanding the current state of metadata creation. this points out the need for further research of current metadata-creation practice. table . electronic mailing lists for the survey electronic mailing lists e-mail address autocat dublin core listserv metadata librarians listserv library and information technology association listserv online audiovisual catalogers electronic discussion list subject authority cooperative program listserv serialist text encoding initiative listserv electronic resources in libraries listserv encoded archival description listserv autocat@listserv.syr.edu dc-libraries@jiscmail.ac.uk metadatalibrarians@lists.monarchos.com lita-l@ala.org olac-list@listserv.acsu.buffalo.edu sacolist@listserv.loc.gov serialst@list.uvm.edu tei-l@listserv.brown.edu eril-l@listserv.binghamton.edu ead@listserv.loc.gov metadata creation practices in digital repositories and collections | park and tosaka and job responsibilities (see table ) clearly show that most of the individuals who completed the survey engage professionally in activities directly relevant to the research objectives, such as descriptive and subject cataloging, metadata creation and management, author- ity control, nonprint and special material cataloging, electronic resource and digital project management, and integrated library system (ils) management. although the largest number of participants ( , or . percent) chose the “other” category regarding their job title (see table ), it is reasonable to assume that the vast majority can be categorized as cataloging and meta- data professionals. most job titles given as “other” are associated with one of the professional activities listed in table . thus it is reasonable to assume that the respondents are in an appropriate position to provide first-hand, accurate information about the current state of metadata creation in their institutions. concerning the institutional background of partici- pants, of the survey participants, fewer than half ( , or . percent) provided institutional information. we believe that this is mostly due to the fact that the question was optional, following a suggestion from the institutional review board at drexel university. of those that provided their institutional background, the majority ( . percent) are from academic libraries, followed by participants from public libraries ( . percent) and from other institutions ( . percent). table . participants’ job responsibilities (multiple responses) job responsibilities number of participants general cataloging (e.g., descriptive and subject cataloging) ( . %) metadata creation and management ( . %) authority control ( . %) nonprint cataloging (e.g., microform, music scores, photographs, video recordings) ( . %) special material cataloging (e.g., rare books, foreign language materials, government documents) ( . %) digital project management ( . %) electronic resource management ( . %) ils management ( . %) other ( . %) survey question: what are your primary job responsibilities? (please check all that apply) table . job titles of participants (multiple responses) job titles number of participants other ( . %) cataloger/cataloging librarian/ catalog librarian ( . %) metadata librarian ( . %) catalog & metadata librarian ( . %) head, cataloging ( . %) electronic resources cataloger ( . %) cataloging coordinator ( . %) head, cataloging & metadata services ( . %) n = . survey question: what is your working job title? (please check all that apply) table . professional activities specified in “other” category in table professional activities number of participants cataloging & metadata creation ( . %) digital projects management ( . %) technical services ( . %) archiving ( . %) electronic resources and serials management ( . %) library system administration/ other ( . %) n = . survey question: if you selected other, please specify. information technology and libraries | september it is noteworthy that use of qualified dc was higher than that of unqualified dc. this result is different from the arl survey and a member survey conducted ■■ results in this section, we will present the findings of this study in the following three areas: ( ) metadata and controlled vocabulary schemata and metadata tools used, ( ) criteria for selecting metadata and controlled vocabulary schemata, and ( ) exposing metadata and metadata guidelines beyond local environments. metadata and controlled vocabulary schemata and metadata tools used a great variety of digital objects were handled by the survey participants, as figure shows. the most frequently han- dled object was text, cited by . percent of the respondents. about three-fourths of the respondents described audiovi- sual materials ( . percent), while . percent described images and . per- cent described archival materials. more than percent of the respondents han- dled electronic resources ( . percent) and digitized resources ( . percent), while approximately half handled born- digital resources ( . percent). the types of materials described in digital collections were diverse, encompassing both digitized and born-digital materi- als; however, digitization accounted for a slightly greater percentage of meta- data creation. to handle these diverse digital objects, the respondents’ institutions employed a wide range of metadata schemata, as figure shows. yet there were a few schemata that were widely used by cataloging and metadata pro- fessionals. specifically, . percent of the respondents’ institutions used marc; dc was also popular, with . percent using unqualified dc and . percent using qualified dc to create metadata. ead also was frequently cited ( . percent). in addition to these major types of metadata schemata, the respondents’ institutions also employed metadata object description schema (mods) ( . percent), visual resource association (vra) core ( . percent), and text encoding initiative (tei) ( . percent). figure . materials/resources handled (multiple responses) survey question: what type of materials/resources do you and your fellow catalogers/metadata librar- ians handle? (please check all that apply) figure . metadata schemata used (multiple responses) survey question: which metadata schema(s) do you and your fellow catalogers/metadata librarians use? (please check all that apply) metadata creation practices in digital repositories and collections | park and tosaka custom metadata elements derives from the imperative to accommodate the perceived needs of local collec- tions and users, as indicated by the two most common responses: ( ) “to reflect the nature of local collec- tions/resources” ( . percent) and ( ) “to reflect the characteristics of target audience/community of local collections” ( . percent). local conditions were also cited from institutional and technical standpoints. many institutions ( . percent) follow existing local practices for cataloging and metadata creation while other insti- tutions ( . percent) are making homegrown metadata additions because of constraints imposed by their local systems. table summarizes the most frequently used con- trolled vocabulary schematas by resource type. by far the most widely used schema across all resource types was lcsh. the preeminence of lcsh evinces the criti- cal role that it plays as the de facto form of controlled vocabulary for subject description. library of congress classification (lcc) was the second choice for all resource types other than images, cultural objects, and archives. for digital collections of these resource types and digitized resources, aat was the second most used controlled vocabulary, a fact that reflects its purpose as a domain-specific terminology used for describing works of art, architecture, visual resources, material culture, and archival materials. while traditional metadata schemata, content stan- dards, and controlled vocabularies such as marc, aacr , and lcsh clearly were preeminent in the majority of the respondents’ institutions, current meta- data creation in digital repositories and collections faces new challenges from the enormous volume of online and digital resources. approximately one-third of the respondents’ institutions ( . percent) were meeting this challenge with tools for semiautomatic metadata generation. yet a majority of respondents ( . percent) indicated that their institutions did not use any such tools for metadata creation and management. this result seems to contrast with ma’s finding that automatic meta- data generation was used in some capacity in nearly by oclc rlg programs (as described in “literature review” on page ). in these surveys, unqualified dc was more frequently cited than qualified dc. one pos- sible explanation of this less frequent use of unqualified dc may lie in the limitations of unqualified dc meta- data semantics. survey respondents also reported on problems using dc metadata, which were mostly caused by semantic ambiguities and semantic overlaps of cer- tain dc metadata elements. limitations and issues of unqualified dc metadata semantics are discussed in depth in park’s study. in light of these results, examin- ing trends of qualified dc use in a future study would be interesting. despite the wide variety of schemata reported in use, there seemed to be an inclination to use only one or two metadata schemata for resource description. as shown in table , the majority of the respondents’ institutions ( . percent) used only one schema for metadata creation, while approximately percent used two or three sche- mata ( . percent and . percent, respectively). the institutions using more than three schemata during the metadata-creation processes comprised only . percent of the respondents. turning to content standards (see figure ), we found that aacr was the most widely used standard, indi- cated by . percent of respondents. this high percentage clearly reflects the continuing preeminence of marc as the metadata schema of choice for digital collections. dc application profiles also showed a large user base, indicated by more than one-third of respondents ( . percent). more than one quarter of the respondents ( . percent) used ead application guidelines as developed by the society of american archivists and the library of congress, while . percent used rlg best practice guidelines for encoded archival description ( ). about one quarter ( . percent) indicated dacs as their content standard. homegrown standards and guidelines are local appli- cation profiles that clarify existing content standards and specify how values for metadata elements are selected and represented to meet the requirements of a particular context. as shown in the results on metadata schemata, it is noteworthy that homegrown content standards and guidelines constituted one of the major choices of partici- pants, indicated by more than one-fifth of the institutions ( . percent). almost two-fifths of the survey partici- pants ( percent) also reported that they add homegrown metadata elements to a given metadata schema. slightly less than half of the participants ( . percent) indicated otherwise. the local practice of creating homegrown content guidelines and metadata elements during the metadata- creation process deserves a separate study; this study only briefly touches on the basis for locally added custom metadata elements. the motivation to create table . number of metadata schemata in use number of metadata schemata in use number of participants ( . %) ( . %) ( . %) or more ( . %) n= . survey question: which metadata schema(s) do you and your fellow catalogers/metadata librarians use the most? (please check all that apply) information technology and libraries | september criteria for selecting metadata and controlled vocabulary schemata what are the factors that have shaped the current state of metadata-creation practices reported thus far? in this section, we turn our attention to constraints that affect decision making at institutions in the selection of meta- data and controlled vocabulary schemata for subject description. figure presents the percentage of different metadata schemata selection criteria described by survey par- ticipants. first, collection-specific considerations clearly played a major role in the selection. the most frequently cited reason was “types of resources” ( . percent). this response reflects the fact that a large number of metadata schemata have been developed, often with wide varia- tion in content and format, to better handle particular two-thirds of arl libraries. because semiautomatic metadata application is reported in-depth in a separate study, we only briefly sketch the topic here. the semiautomatic metadata application tools used in the respondents’ digital reposi- tories and collections can be classified into five categories of common characteristics: ( ) metadata format conver- sion, ( ) templates and editors for metadata creation, ( ) automatic metadata creation, ( ) library system for bibliographic and authority control, and ( ) metadata harvesting and importing tools. as table illustrates, among those institutions that have introduced semiautomatic metadata generation tools, “metadata format conversion” ( . percent) and “templates and editors for metadata creation” ( per- cent) are the two most frequently cited tools. figure . content standards used (multiple responses) survey question: what content standard(s) and/or guidelines do you and your fellow catalogers/metadata librarians use? (please check all that apply) metadata creation practices in digital repositories and collections | park and tosaka job responsibility, “expertise of staff” ( . percent) and “integrated library system” ( . percent) appeared to highlight the key role that marc continues to play in the metadata-creation process for digital collections (see fig- ure ). “budget” also appeared to be an important factor in metadata selection ( . percent), showing that funding levels played a considerable role in metadata decisions. types of information resources. the primary factor in selecting metadata schemata is their suit- ability for describing the most common type of resources han- dled by the survey participants. the second and third most common criteria, “target users/ audience” ( . percent) and “subject matters of resources” ( . percent), also seem to reflect how domain-specific metadata schemata are applied. in making decisions on meta- data schemata, respondents weighed materials in particular subject areas (e.g., art, education, and geography) and the needs of particu- lar communities of practice as their primary users and audiences. however, existing technological infrastructure and resource constraints also determine options. given the prominence of general library cataloging as a primary table . the most frequently used controlled vocabulary schema(s) by resource type (multiple responses) lcsh lcc ddc aat tgm ulan tgn other text . % ( ) . % ( ) . % ( ) . % ( ) . % ( ) . % ( ) . % ( ) . % ( ) audiovisual materials . % ( ) . % ( ) . % ( ) . % ( ) . % ( ) . % ( ) . % ( ) . % ( ) cartographic materials . % ( ) . % ( ) . % ( ) . % ( ) . % ( ) . % ( ) . % ( ) . % ( ) images . % ( ) . % ( ) . % ( ) . % ( ) . % ( ) . % ( ) . % ( ) . % ( ) cultural objects (e.g., museum objects) . % ( ) . % ( ) . % ( ) . % ( ) . % ( ) . % ( ) . % ( ) . % ( ) archives . % ( ) . % ( ) . % ( ) . % ( ) . % ( ) . % ( ) . % ( ) . % ( ) electronic resources . % ( ) . % ( ) . % ( ) . % ( ) . % ( ) . % ( ) . % ( ) . % ( ) digitized resources . % ( ) . % ( ) . % ( ) . % ( ) . % ( ) . % ( ) . % ( ) . % ( ) born-digital resources . % ( ) . % ( ) . % ( ) . % ( ) . % ( ) . % ( ) . % ( ) . % ( ) survey question: which controlled vocabulary schema(s) do you and your fellow catalogers/metadata librarians use most? (please check all that apply) table . types of semi-automatic metadata generation tools in use types response rating metadata format conversion ( . %) templates and editors for metadata creation ( . %) automatic metadata creation ( . %) library system for bibliographic and authority control ( . %) metadata harvesting and importing tools ( . %) n = . survey question: please describe the (semi)automatic metadata generation tools you use. information technology and libraries | september the software used by their insti- tutions—i.e., “integrated library system” ( . percent), “digital collection or asset management software” ( . percent), “institu- tional repository software” ( . percent), “union catalogs” ( . percent), and “archival manage- ment software” ( . percent)—as a reason for their selection of meta- data schemata. metadata decisions thus seem to be driven by a vari- ety of local technology choices for developing digital repositories and collections. as shown in figure , similar patterns are observed with regard to selection criteria for controlled vocabulary schemata. three of the four selection criteria receiv- ing majority responses—“target users/audience” ( . percent), “type of resources” ( . percent), and “nature of the collection” ( . percent)—suggest that controlled vocabulary decisions are influ- enced primarily by the substantive purpose and scope of controlled vocabularies for local collections. a major consideration seems to be whether particular controlled vocabularies are suitable for rep- resenting standard data values to improve access and retrieval for target audiences. “metadata standards,” another selection criteria frequently cited in the survey ( . percent), reflects how some domain-spe- cific metadata schemata tend to dictate the use of particular con- trolled vocabularies. at the same time, the results also suggest that resources and technological infra- structure available to institutions were also important reasons for their selections. “expertise of staff” ( . percent) seems to be a straightforward practical rea- son: the application of controlled vocabularies is highly dependent on the width and depth of staff expertise available. likewise, when implementing controlled vocabularies in the digital environment, some institutions also took into account at the same time, it is noteworthy that while responses were not mutually exclusive, many respondents cited figure . criteria for selecting metadata schemata (multiple responses) question: which criteria were applied in selecting metadata schemata? (please check all that apply) figure . criteria for selecting controlled vocabulary schemata (multiple responses) question: which criteria are applied in selecting controlled vocabulary schemata? (please check all that apply) metadata creation practices in digital repositories and collections | park and tosaka for search engines and . percent for oai harvesters), a result that may be interpreted as a tendency to create metadata primarily for local audiences. why do many institutions fail to make their locally created meta- data available to other institutions despite wide consensus on the importance of metadata sharing in a networked world? responses from those institutions exposing none or not all of their metadata (see table ) reveal that financial, personnel, and technical issues are major hin- drances in promoting the exposure of metadata outside the immediate local environment. some institutions are not confident that their current metadata practices are able to sat- isfy the technical requirements for producing standards-based interop- erable metadata. another reason frequently mentioned is copyright concerns about limited-access mate- rials. yet some respondents simply do not see any merit to exposing their item-level metadata, citing its relative uselessness for resource discovery outside their institutions. as stated earlier, the practice of adding home- grown metadata elements seems common among many institutions. while locally created metadata elements accommodate local needs and requirements, they may also hinder metadata interoperability across digital repositories and collections if mechanisms for finding information about such locally defined extensions and variants are absent. homegrown metadata guidelines document local data models and function as an essential mechanism for metadata creation and quality assurance within and across digital repositories and collections. in this regard, it is essential to examine locally created metadata guidelines and best practices. however, the results of the survey analysis evince that the vast majority of institutions ( . percent) provided no public access to local application profiles on their websites while only . percent of respondents’ institutions made them available online to the public. ■■ conclusion metadata plays an essential role in managing, organizing, and searching for information resources. in the networked existing system features for authority control and con- trolled vocabulary searching, as exhibited by . percent of responses for “digital collection/or asset management software.” exposing metadata and metadata guidelines beyond local environments metadata interoperability across distributed digital repositories and collections is fast becoming a major issue. the proliferation of open-source and commercial digital library platforms using a variety of metadata sche- mata has implications on the librarians’ ability to create shareable and interoperable metadata beyond the local environment. to what extent are mechanisms for sharing metadata integrated into the current metadata-creation practices described by the respondents? figure summarizes the responses concerning the uses of three major mechanisms for metadata exposure. approximately half of respondents exposed at least some of their metadata to search engines ( . percent) and union catalogs such as oclc worldcat ( . percent). more than one-third of the respondents exposed all or some of their metadata through oai harvesters ( . percent). about half or more of the respondents either did not expose their metadata or were not sure about the current operations at their institutions (e.g., . percent figure . mechanism to expose metadata (multiple responses) survey question: do you/your organization expose your metadata to oai (open archives initiative) harvesters, union catalogs or search engines? information technology and libraries | september the dc metadata schema is the second most widely employed according to this study, with qualified dc used by . percent of responding institutions and unqualified dc used by . percent. ead is another fre- quently cited schema ( . percent), followed by mods ( . percent), vra ( . percent), and tei ( . percent). a trend of qualified dc being used ( . percent) more often than unqualified dc ( . percent) is noteworthy. one possible explanation of this trend may be derived from the fact that semantic ambiguities and overlaps in some of the unqualified dc elements interfere with use in resource description. given the earlier surveys report- ing the higher use of unqualified dc over qualified dc, more in-depth examination of their use trends may be an important avenue for future studies. despite active research and promising results obtained from some experimental tools, practical applications of semiautomatic metadata generation have been incor- porated into the metadata-creation processes by only one-third of survey participants. the leading criteria in selecting metadata and controlled vocabulary schemata are derived from collec- tion-specific considerations of the type of resources, the nature of the collections, and the needs of primary users and communities. existing technological infrastructure, encompassing digital collection or asset management software, archival management software, institutional repository software, integrated library systems, and union catalogs also greatly influence the selection pro- cess. the skills and knowledge of metadata professionals and the expertise of staff also are significant factors in understanding current practices in the use of metadata schemata and controlled vocabularies for subject access across distributed digital repositories and collections. the survey responses reveal that metadata interop- erability remains a challenge in the current networked environment despite growing awareness of its impor- tance. for half of the survey respondents, exposing metadata to the service providers, such as oai harvesters, union catalogs, and search engines, does not seem to be a high priority because of local financial, personnel, and technical constraints. locally created metadata elements are added in many digital repositories and collections in large part to meet local descriptive needs and serve the target user community. while locally created metadata elements accommodate local needs, they may also hinder metadata interoperability across digital repositories and collections when shareable mechanisms are not in place for such locally defined extensions and variants. locally created metadata guidelines and application profiles are essential for metadata creation and quality assurance; however, most custom content guidelines and best practices ( percent) are not made publicly avail- able. the lack of a mechanism to facilitate public access to local application profiles and metadata guidelines may environment, the enormous volume of online and digital resources creates an impending research need to evalu- ate the issues surrounding the metadata-creation process and the employment of controlled vocabulary schemata across ever-growing distributed digital repositories and collections. in this paper we explored the current status of metadata-creation practices through an examination of survey responses drawn mostly from cataloging and metadata professionals (see tables , , and ). the results of the study indicate that current metadata practices still do not create conditions for interoperability. despite the proliferation of newer metadata sche- mata, the survey responses showed that marc currently remains the most widely used schema for providing resource description and access in digital repositories, collections, and libraries. the continuing predominance of marc goes hand-in-hand with the use of aacr as the primary content standard for selecting and represent- ing data values for descriptive metadata elements. lcsh is used as the de facto controlled vocabulary schema for providing subject access in all types of digital repositories and collections, while domain-specific subject terminolo- gies such as aat are applied at significantly higher rates in digital repositories handling nonprint resources such as images, cultural objects, and archival materials. table . sample reasons for not exposing metadata not all our metadata conforms to standards required not all our metadata is oai compliant lack of expertise and time and money to develop it it restrictions security concerns on the part of our information technology department some collections/records are limited access and not open to the general public we think that having worldcat available for traditional library materials that many libraries have is a better service to people than having each library dump our catalog out on the web varies by tool and collection, but usually a restriction on the material, a technical barrier, or a feeling that for some collections the data is not yet sufficiently robust “still in a work in progress” survey question: if you selected “some, but not all” or “no” in question [see figure ], please tell why you do not expose your metadata. metadata creation practices in digital repositories and collections | park and tosaka presented at dublin core conference: supporting commu- nities of discourse and practice—metadata research & appli- cations, seattle, wash., sept. –oct. , ), http://dcpapers .dublincore.org/ojs/pubs/article/view/ / (accessed mar. , ); sarah currier et al., “quality assurance for digital learning object repositories: issues for the metadata-creation process,” alt-j ( ): – . . jin ma, metadata, spec kit (washington, d.c.: asso- ciation of research libraries, ): , . . ibid., , – . . karen smith-yoshimura, rlg programs descriptive meta- data practices survey results (dublin, ohio: oclc, ): – , http://www.oclc.org/programs/publications/reports/ - .pdf (accessed mar. , ); karen smith-yoshimura and diane cellentani, rlg programs descriptive metadata practices survey results: data supplement (dublin, ohio: oclc, ): , http://www.oclc.org/programs/publications/reports/ - .pdf (accessed mar. , ). . carole palmer, oksana zavalina, and megan mustafoff, “trends in metadata practices: a longitudinal study of col- lection federation” (paper presented at the seventh acm/ iees-cs joint conference on digital libraries, vancouver, brit- ish columbia, canada, june – , ), http://hdl.handle .net/ / (accessed mar. , ). . smith-yoshimura, rlg programs descriptive metadata prac- tices survey results, ; smith-yoshimura and cellentani, rlg programs descriptive metadata practices survey results, . . ma, metadata, , – . . smith-yoshimura, rlg programs descriptive metadata prac- tices survey results, ; smith-yoshimura and cellentani, rlg programs descriptive metadata practices survey results, – . . karen markey et al., census of institutional repositories in the united states: miracle project research findings (washing- ton, d.c.: council on library & information resources, ): , – , http://www.clir.org/pubs/reports/pub /pub .pdf (accessed mar. , ). . yoshimura and cellentani, rlg programs descriptive meta- data practices survey results, . . university of houston libraries institutional repository task force, institutional repositories, spec kit (washington, d.c.: association of research libraries, ): , . . ma, metadata, , . . smith-yoshimura, rlg programs descriptive metadata prac- tices survey results, , ; smith-yoshimura and cellentani, rlg programs descriptive metadata practices survey results, – . . for the metrics of job responsibilities used to analyze job descriptions and competencies of cataloging and metadata professionals, see jung-ran park, caimei lu, and linda marion, “cataloging professionals in the digital environment: a content analysis of job descriptions,” journal of the american society for information science & technology ( ): – ; jung-ran park and caimei lu, “metadata professionals: roles and competen- cies as reflected in job announcements, – ,” cataloging & classification quarterly ( ): – . . ma, metadata; smith-yoshimura, rlg programs descriptive metadata practices survey result. . jung-ran park and eric childress, “dublin core metadata semantics: an analysis of the perspectives of information pro- fessionals,” joural of information science , no. ( ): – . . park, “semantic interoperability.” . jung-ran park, “metadata quality in digital repositories: hinder cross-checking for quality metadata and creating shareable metadata that can be harvested for a high level of consistency and interoperability across distributed digital collections and repositories. development of a searchable registry for publicly available metadata guide- lines has potential to enhance metadata interoperability. a constraining factor of this study derives from the participant population; thus we have not attempted to generalize the findings of the study. however, results indicate a pressing need for a common data model that is shareable and interoperable across ever-growing dis- tributed digital repositories and collections. development of such a common data model demands future research of a practical and interoperable mediation mechanism underlying local implementation of metadata elements, semantics, content standards, and controlled vocabular- ies in a world where metadata can be distributed and shared widely beyond the immediate local environment and user community. (other issues such as semiauto- matic metadata application, dc metadata semantics, custom metadata elements, and the professional devel- opment of cataloging and metadata professionals are explained in-depth in separate studies.) for future studies, incorporation of other research methods (such as follow-up telephone surveys and face-to-face focus group interviews) could be used to better understand the cur- rent status of metadata-creation practices. institutional variation also needs be taken into account in the design of future studies. ■■ acknowledgments this study is supported through an early career develop- ment research award from the institute of museum and library services. we would like to express our apprecia- tion to the reviewers for their invaluable comments. references . jung-ran park, “semantic interoperability and metadata quality: an analysis of metadata item records of digital image collections,” knowledge organization ( ): – ; rachel heery, “metadata futures: steps toward semantic interoper- ability,” in metadata in practice, ed. diane i. hillman and elaine l. westbrooks, – (chicago: ala, ); jung-ran park, “semantic interoperability across digital image collections: a pilot study on metadata mapping” (paper presented at the canadian association for information science annual con- ference, london, ontario, june – , ), http://www.cais-acsi .ca/proceedings/ /park_j_ .pdf (accessed mar. , ). . jane barton, sarah currier, and jessie m. n. hey, “building quality assurance into metadata creation: an analysis based on the learning objects and e-prints communities of practice” (paper information technology and libraries | september a survey of the current state of the art,” in “metadata and open access repositories,” ed. michael s. babinec and holly mercer, special issue, cataloging & classification quarterly , no. / ( ): – . . ma, metadata, , . the oclc rlg survey found that about percent of the respondents were able to generate some metadata automatically. see smith-yoshimura, rlg programs descriptive metadata practices survey results, ; yoshimura and cellentani, rlg programs descriptive metadata practices survey results, . . jung-ran park and caimei lu, “application of semi- automatic metadata generation in libraries: types, tools, and techniques,” library & information science research , no. ( ): – . . park, “semantic interoperability”; sarah l. shreeves et al., “is ‘quality’ metadata ‘shareable’ metadata? the implications of local metadata practices for federated collections” (paper presented at the th national conference of the association of college and research libraries, apr. – , , minneapolis, minnesota), https://www.ideals.uiuc.edu/handle/ / (accessed mar. , ); amy s. jackson et al., “dublin core metadata harvested through oai-pmh,” journal of library meta- data , no. ( ): – ; lois mai chan and marcia lei zeng, “metadata interoperability and standardization—a study of methodology part i: achieving interoperability at the schema level,” d-lib magazine , no. ( ), http://www.dlib.org/ dlib/june /chan/ chan.html (accessed mar. , ); marcia lei zeng and lois mai chan, “metadata interoperability and standardization—a study of methodology part ii: achieving interoperability at the record and repository levels,” d-lib magazine , no. ( ), http://www.dlib.org/dlib/june / zeng/ zeng.html (accessed mar. , ). . thomas r. bruce and diane i. hillmann, “the con- tinuum of metadata quality: defining, expressing, exploiting,” in metadata in practice, ed. hillman and westbrooks, – ; heery, “metadata futures”; park, “metadata quality in digital repositories.” . jung-ran park, ed., “metadata best practices: current issues and future trends,” special issue, journal of library meta- data , no. / ( ). . see park, “semantic interoperability”; park and childress, “dublin core metadata semantics.” . park and childress, “dublin core metadata semantics”; park and lu, “application of semi-automatic metadata genera- tion in libraries.” batch loading collections into dspace | walsh maureen p. walsh batch loading collections into dspace: using perl scripts for automation and quality control colleagues briefly described batch loading marc meta- data crosswalked to dspace dublin core (dc) in a poster session. mishra and others developed a perl script to create the dspace archive directory for batch import of electronic theses and dissertations (etds) extracted with a java program from an in-house bibliographic database. mundle used perl scripts to batch process etds for import into dspace with marc catalog records or excel spreadsheets as the source metadata. brownlee used python scripts to batch process comma-separated values (csv) files exported from filemaker database software for ingest via the dspace item importer. more in-depth descriptions of batch loading are pro- vided by thomas; kim, dong, and durden; proudfoot et al.; witt and newton; drysdale; ribaric; floyd; and averkamp and lee. however, irrespective of reposi- tory software, each describes a process to populate their repositories dissimilar to the workflows developed for the knowledge bank in approach or source data. thomas describes the perl scripts used to convert marc catalog records into dc and to create the archive directory for dspace batch import. kim, dong, and durden used perl scripts to semiauto- mate the preparation of files for batch loading a university of texas harry ransom humanities research center (hrc) collection into dspace. the xml source metadata they used was generated by the national library of new zealand metadata extraction tool. two subsequent proj- ects for the hrc revisited the workflow described by kim, dong, and durden. proudfoot and her colleagues discuss importing meta- data-only records from departmental refbase, thomson reuters endnote, and microsoft access databases into eprints. they also describe an experimental perl script written to scrape lists of publications from personal web- sites to populate eprints. two additional workflow examples used citation databases as the data source for batch loading into repositories. witt and newton provide a tutorial on trans- forming endnote metadata for digital commons with xslt (extensible stylesheet language transformations). drysdale describes the perl scripts used to convert thomson reuters reference manager files into xml for the batch loading of metadata-only records into the university of glascow’s eprints repository. the glascow eprints batch workflow is additionally described by robertson and nixon and greig. several workflows were designed for batch loading etds into repositories. ribaric describes the automatic this paper describes batch loading workflows developed for the knowledge bank, the ohio state university’s institutional repository. in the five years since the incep- tion of the repository approximately percent of the items added to the knowledge bank, a dspace repository, have been batch loaded. most of the batch loads utilized perl scripts to automate the process of importing meta- data and content files. custom perl scripts were used to migrate data from spreadsheets or comma-separated values files into the dspace archive directory format, to build collections and tables of contents, and to provide data quality control. two projects are described to illus- trate the process and workflows. t he mission of the knowledge bank, the ohio state university’s (osu) institutional repository, is to col- lect, preserve, and distribute the digital intellectual output of osu’s faculty, staff, and students. the staff working with the knowledge bank have sought from its inception to be as efficient as possible in adding content to dspace. using batch loading workflows to populate the repository has been integral to that efficiency. the first batch load into the knowledge bank was august , . over the next four years, collections con- taining , items were batch loaded, representing percent of the items and percent of the collections in the knowledge bank. these batch loaded collections vary from journal issues to photo albums. the items include articles, images, abstracts, and transcripts. the majority of the batch loads, including the first, used custom perl scripts to migrate data from microsoft excel spreadsheets into the dspace batch import format for descriptive meta- data and content files. perl scripts have been used for data cleanup and quality control as part of the batch load pro- cess. perl scripts, in combination with shell scripts, have also been used to build collections and tables of contents in the knowledge bank. the workflows using perl scripts to automate batch import into dspace have evolved through an iterative process of continual refinement and improvement. two knowledge bank projects are pre- sented as case studies to illustrate a successful approach that may be applicable to other institutional repositories. ■■ literature review batch ingesting is acknowledged in the literature as a means of populating institutional repositories. there are examples of specific batch loading processes mini- mally discussed in the literature. branschofsky and her maureen p. walsh (walsh. @osu.edu) is metadata librarian/ assistant professor, the ohio state university libraries, colum- bus, ohio. information technology and libraries | september relational database postgresql . . on the red hat enterprise linux operating system. the structure of the knowledge bank follows the hierarchical arrangement of dspace. communities are at the highest level and can be divided into subcommunities. each community or subcommunity contains one or more collections. all items—the basic archival elements in dspace—are con- tained within collections. items consist of metadata and bundles of bitstreams (files). dspace supports two user interfaces: the original interface based on javaserver pages (jspui) and the newer manakin (xmlui) interface based on the apache cocoon framework. at this writing, the knowledge bank continues to use the jspui interface. the default metadata used by dspace is a qualified dc schema derived from the dc library application profile. the knowledge bank uses a locally defined extended version of the default dspace qualified dc schema, which includes several additional element quali- fiers. the metadata management for the knowledge bank is guided by a knowledge bank application profile and a core element set for each collection within the reposi- tory derived from the application profile. the metadata librarians at osul create the collection core element sets in consultation with the community representatives. the core element sets serve as metadata guidelines for sub- mitting items to the knowledge bank regardless of the method of ingest. the primary means of adding items to collections in dspace, and the two ways used for knowledge bank ingest, are ( ) direct (or intermediated) author entry via the dspace web item submission user inter- face and ( ) in batch via the dspace item importer. recent enhancements to dspace, not yet fully explored for use with the knowledge bank, include new ingest options using simple web-service offering repository deposit (sword), open archives initiative object reuse and exchange (oai-ore), and dspace package import- ers such as the metadata encoding and transmission standard submission information package (mets sip) preparation of etds from the internet archive (http:// www.archive.org/) for ingest into dspace using php utilities. floyd describes the processor developed to automate the ingest of proquest etds via the dspace item importer. also using proquest etds as the source data, averkamp and lee described using xslt to transform the proquest data to bepress’ (the berkeley electronic press) schema for batch loading into a digital commons repository. the knowledge bank workflows described in this paper use perl scripts to generate dc xml and create the archive directory for batch loading metadata records and content files into dspace using excel spreadsheets or csv files as the source metadata. ■■ background the knowledge bank, a joint initiative of the osu libraries (osul) and the osu office of the chief information officer, was first registered in the registry of open access repositories (roar) on september , . as of december the repository held , items in , collections. the knowledge bank uses dspace, the open-source java-based repository software jointly developed by the massachusetts institute of technology libraries and hewlett-packard. as a dspace reposi- tory, the knowledge bank is organized by communities. the fifty-two communities currently in the knowledge bank include administrative units, colleges, departments, journals, library special collections, research centers, symposiums, and undergraduate honors theses. the com- monality of the varied knowledge bank communities is their affiliation with osu and their production of knowl- edge in a digital format that they wish to store, preserve, and distribute. the staff working with the knowledge bank includes a team of people from three osul areas—technical services, information technology, and preservation—and the contracted hours of one systems developer from the osu office of information technology (oit). the osul team members are not individually assigned full-time to the repository. the current osul team includes a librarian reposi- tory manager, two metadata librarians, one systems librarian, one systems developer, two technical services staff members, one preservation staff mem- ber, and one graduate assistant. the knowledge bank is cur- rently running dspace . . and the figure . dspace simple archive format archive_directory/ item_ / dublin_core.xml--qualified dublin core metadata contents --text file containing one line per filename file_l.pdf --files to be added as bitstreams to the item file_ .pdf item_ / dublin_core.xml file_ .pdf ... batch loading collections into dspace | walsh ■■ case studies the issues of the ohio journal of science ojs was jointly published by osu and the ohio academy of science (oas) until , when oas took over sole control of the journal. the issues of ojs are archived in the knowledge bank with a two year rolling wall embargo. the issues for through , a total of issues containing , articles, were batch loaded into the knowledge bank. due to rights issues, the retrospec- tive batch loading project had two phases. the project to digitize ojs began with the – issues that osu had the rights to digitize and make publicly available. osu later acquired the rights for –present, and (accounting for the embargo period) – became phase of the project. the two phases of batch loads were the most complicated automated batch loading processes developed to date for the knowledge bank. to batch load phase in and phase in , the systems devel- opers working with the knowledge bank wrote scripts to build collections, generate dc xml from the source metadata, create the archive directory, load the metadata and content files, create tables of contents, and load the tables of contents into dspace. the ojs community in the knowledge bank is orga- nized by collections representing each issue of the journal. the systems developers used scripts to automate the building of the collections in dspace because of the number needed as part of the retrospective project. the individual articles within the issues are items within the collections. there is a table of contents for the articles in each issue as part of the collection homepages. again, due to the number required for the retrospective project, the systems developers used scripts to automate the cre- ation and loading of the tables of contents. the tables of contents are contained in the html introductory text sec- tion of the collection pages. the tables of contents list title, authors, and pages. they also include a link to the item record and a direct link to the article pdf that includes the file size. for each phase of the ojs project, a vendor con- tracted by osul supplied the article pdfs and an excel spreadsheet with the article-level metadata. the metadata format. this paper describes ingest via the dspace batch item importer. the dspace item importer is a command-line tool for batch ingesting items. the importer uses a simple archive format diagramed in figure . the archive is a directory of items that contain a subdirectory of item metadata, item files, and a contents file listing the bitstream file names. each item’s descriptive metadata is contained in a dc xml file. the format used by dspace for the dc xml files is illustrated in figure . automating the process of creating the unix archive directory has been the main function of the perl scripts written for the knowledge bank batch loading workflows. a systems developer uses the test mode of the dspace item importer tool to validate the item directories before doing a batch load. any significant errors are corrected and the process is repeated. after a successful test, the batch is loaded into the staging instance of the knowledge bank and quality checked by a metadata librarian to identify any unexpected results and script or data problems that need to be corrected. after a successful load into the staging instance the batch is loaded into the production instance of the knowledge bank. most of the knowledge bank batch loading work- flows use excel spreadsheets or csv files as the source for the descriptive item metadata. the creation of the metadata contained in the spreadsheets or files has var- ied by project. in some cases the metadata is created by osul staff. in other cases the metadata is supplied by knowledge bank communities in consultation with a metadata librarian or by a vendor contracted by osul. whether the source metadata is created in-house or exter- nally supplied, osul staff are involved in the quality control of the metadata. several of the first communities to join the knowledge bank had very large retrospective collection sets to archive. the collection sets of two of those early adopt- ers, the journal issues of the ohio journal of science (ojs) and the abstracts of the osu international symposium on molecular spectroscopy currently account for percent of the items in the knowledge bank. the successful batch loading workflows developed for these two com- munities—which continue to be active content suppliers to the repository—are presented as case studies. figure . dspace qualified dublin core xml notes on the bird life of cedar point - griggs, robert f. information technology and libraries | september article-level metadata to knowledge bank dc, as illus- trated in table . the systems developers used the mapping as a guide to write perl scripts to transform the vendor metadata into the dspace schema of dc. the workflow for the two phases was nearly identical, except each phase had its own batch loading scripts. due to a staff change between the two phases of the project, a former osul systems developer was responsible for batch loading phase and the oit systems developer was responsible for phase . the phase scripts were all writ- ten in perl. the four scripts written for phase created the archive directory, performed database operations to build the collections, generated the html introduction table of contents for each collection, and loaded the tables of contents into dspace via the database. for phase , the oit systems developer modified and added to the phase batch processing scripts. this case study focuses on phase of the project. batch processing for phase of ojs the annotated scripts the oit systems developer used for phase of the ojs project are included in appen- dix a, available on the italica weblog (http://ital-ica .blogspot.com/). a shell script (mkcol.sh) added collec- tions based on a listing of the journal issues. the script performed a login as a selected user id to the dspace web interface using the web access tool curl. a subsequent simple looping perl script (mkallcol.pl) used the stored credentials to submit data via this channel to build the collections in the knowledge bank. the metadata.pl script created the archive directory for each collection. the oit systems developer added the pdf file for each item to unix. the vendor-supplied meta- data was saved as unicode text format and transferred to unix for further processing. the developer used vi com- mands to manually modify metadata for characters illegal in xml (e.g., “<” and “&”). (although manual steps were used for this project, the oit systems developer improved the perl scripts for subsequent projects by add- ing code for automated transformation of the input data to help ensure xml validity.) the metadata.pl script then processed each line of the metadata along with the cor- responding data file. for each item, the script created the dc xml file and the contents file and moved them and the pdf file to the proper directory. load sets for each col- lection (issue) were placed in their own subdirectory, and a load was done for each subdirectory. the items for each collection were loaded by a small perl script (loaditems. pl) that used the list of issues and their collection ids and called a shell script (import.sh) for the actual load. the tables of contents for the issues were added to the knowledge bank after the items were loaded. a perl script (intro.pl) created the tables of contents using the meta- data and the dspace map file, a stored mapping of item received from the vendor had not been customized for the knowledge bank. the ojs issues were sent to a vendor for digitization and metadata creation before the knowledge bank was chosen as the hosting site of the digitized jour- nal. the osu digital initiatives steering committee proposal for the ojs digitization project had predated the knowledge bank dspace instance. osul staff performed quality-control checks of the vendor-supplied metadata and standardized the author names. the vendor supplied the author names as they appeared in the articles—in direct order, comma separated, and including any “and” that appeared. in addition to other quality checks per- formed, osul staff edited the author names in the spreadsheet to conform to dspace author-entry conven- tion (surname first). semicolons were added to separate author names, and the extraneous ands were removed. a former metadata librarian mapped the vendor-supplied table . mapping of vendor metadata to qualified dublin core vendor-supplied metadata knowledge bank dublin core file [n/a: pdf file name] cover title dc.identifier.citation* issn dc.identifier.issn vol. dc.identifier.citation* iss. dc.identifier.citation* cover date dc.identifier.citation* year dc.date.issued month dc.date.issued fpage dc.identifier.citation* lpage dc.identifier.citation* article title dc.title author names dc.creator institution dc.description abstract dc.description.abstract n/a dc.language.iso n/a dc.rights n/a dc.type *format: [cover title]. v[vol.], n[iss.] ([cover date]), [fpage]-[lpage] batch loading collections into dspace | walsh directories to item handles created during the load. the tables of contents were added to the knowledge bank using a shell script (installintro.sh) similar to what was used to create the collections. installintro.sh used curl to simulate a user adding the data to dspace by performing a login as a selected user id to the dspace web interface. a simple looping perl script (ldallintro.pl) called installintro.sh and used the stored credentials to submit the data for the tables of contents. the abstracts of the osu international symposium on molecular spectroscopy the knowledge bank contains the abstracts of the papers presented at the osu international symposium on molecular spectroscopy (mss), which has met annually since . beginning with the symposium, the complete presentations from authors who have autho- rized their inclusion are archived along with the abstracts. the mss community in the knowledge bank currently contains , items grouped by decade into six col- lections. the six collections were created “manually” via the dspace web interface prior to the batch loading of the items. the retrospective years of the symposium ( – ) were batch loaded in three phases in . each symposium year following the retrospective loads was batch loaded individually. retrospective mss batch loads the majority of the abstracts for the retrospective loads were digitized by osul. a vendor was contracted by osul to digitize the remainder and to supply the meta- data for the retrospective batch loads. the files digitized by osul were sent to the vendor for metadata capture. osul provided the vendor a metadata template derived from the mss core element set. the metadata taken from the abstracts comprised author, affiliation, title, year, session number, sponsorship (if applicable), and a full transcription of the abstract. to facilitate searching, the formulas and special characters appearing in the titles and abstracts were encoded using latex, a document prepara- tion system used for scientific data. the vendor delivered the metadata in excel spreadsheets as per the spreadsheet template provided by osul. quality-checking the meta- data was an essential step in the workflow for osul. the metadata received for the project required revisions and data cleanup. the vendor originally supplied incomplete files and spreadsheets that contained data errors, includ- ing incorrect numbering, data in the wrong fields, and inconsistency with the latex encoding. the three knowledge bank batch load phases for the retrospective mss project corresponded to the staged receipt of metadata and digitized files from the vendor. the annotated scripts used for phase of the project, which included twenty years of the osu international symposium between and , are included in appendix b, available on the italica weblog. the oit systems developer saved the metadata as a tab-separated file and added it to unix along with the abstract files. a perl script (mkxml .pl) transformed the metadata into dc xml and created the archive directories for load- ing the metadata and abstract files into the knowledge bank. the script divided the directories into separate load sets for each of the six collections and accounted for the inconsistent naming of the abstract files. the script added the constant data for type and language that was not included in the vendor-supplied metadata. unlike the ojs project, where multiple authors were on the same line of the metadata file, the mss phase script had to code for authors and their affiliations on separate lines. once the load sets were made, the oit systems devel- oper ran a shell script to load them. the script (import_ collections.sh) was used to run the load for each set so that the dspace item import command did not need to be constructed each time. annual mss batch loads a new workflow was developed for batch loading the annual mss collection additions. the metadata and item files for the annual collection additions are supplied by the mss community. the community provides the symposium metadata in a csv file and the item files in a tar archive file. the symposium uses a web form for latex–formatted abstract submissions. the community processes the electronic symposium submissions with a perl script to create the csv file. the metadata delivered in the csv file is based on the template created by the author, which details the metadata requirements for the project. the oit systems developer borrowed from and modi- fied earlier perl scripts to create a new script for batch processing the metadata and files for the annual symposium collection additions. to assist with the development of the new script, i provided the developer a mapping of the community csv headings to the knowledge bank dc fields. i also provided a sample dc xml file to illustrate the desired result of the perl transformation of the com- munity metadata into dc xml. for each new year of the symposium, i create a sample dc xml result for an item to check the accuracy of the script. a dc xml example from a mss item is included in appendix c, available on the italica weblog. unlike the previous retrospective mss loads in which the script processed multiple years of the symposium, the new script processes one year at a time. the annual symposiums are batch loaded indi- vidually into one existing mss decade collection. the new script for the annual loads was tested and refined by load- ing the symposium into the staging instance of the information technology and libraries | september ■■ summary and conclusion each of the batch loads that used perl scripts had its own unique features. the format of content and associ- ated metadata varied considerably, and custom scripts to convert the content and metadata into the dspace import format were created on a case-by-case basis. the differ- ences between batch loads included the delivery format of the metadata, the fields of metadata supplied, how metadata values were delimited, the character set used for the metadata, the data used to uniquely identify the files to be loaded, and how repeating metadata fields were identi- fied. because of the differences in supplied metadata, a separate perl script for generating the dc xml and archive directory for batch loading was written for each project. each new perl script borrowed from and modified earlier scripts. many of the early batch loads were firsts for the knowledge bank and the staff working with the reposi- tory, both in terms of content and in terms of metadata. dealing with community- and vendor-supplied metadata and various encodings (including latex), each of the early loads encountered different data obstacles, and in each case solutions were written in perl. the batch loading code has matured over time, and the progression of improvements is evident in the example scripts included in the appendixes. batch loading can greatly reduce the time it takes to add content and metadata to a repository, but successful knowledge bank. problems encountered with character encoding and file types were resolved by modifying the script. the metadata and files for the symposium years , , and were made available to osul in , and each year was individually loaded into the existing knowledge bank col- lection for that decade. these first three years of community-supplied csv files contained author metadata inconsistent with knowledge bank author entries. the names were in direct order, upper- case, split by either a semicolon or “and,” and included extraneous data, such as an address. the oit systems developer wrote a perl script to correct the author metadata as part of the batch loading workflow. an annotated section of that script illustrating the author modifica- tions is included in appendix d, available on the italica weblog. the mss com- munity revised the perl script they used to generate the csv files by including an edited version of this author entry cor- rection script and were able to provide the expected author data for and . the author entries received for these years were in inverted order (surname first) and mixed case, were semicolon separated, and included no extraneous data. the receipt of consistent data from the community for the last two years has facilitated the stan- dardized workflow for the annual mss loads. the scripts used to batch load the symposium year are included in appendix e, which appears at the end of this text. the oit systems developer unpacked the tar file of abstracts and presentations into a directory named for the year of the symposium on unix. the perl script written for the annual mss loads (mkxml. pl) was saved on unix and renamed mkxml .pl. the script was edited for (including the name of the csv file and the location of the directories for the unpacked files and generated xml). the csv headings used by the community in the new file were checked and verified against the extract list in the script. once the perl script was up-to-date and the base directory was created, the oit systems developer ran the perl script to gener- ate the archive directory set for import. the import.sh script was then edited for and run to import the new symposium year into the staging instance of the knowledge bank as a quality check prior to loading into the live repository. the brief item view of an example mss item archived in the knowledge bank is shown in figure . figure . mss archived item example batch loading collections into dspace | walsh proceedings of the international conference on dublin core and metadata applications: supporting com- munities of discourse and practice—metadata research & applications, seattle, washington, , http://dcpapers .dublincore.org/ojs/pubs/article/view/ / (accessed dec. , ). . r. mishra et al., “development of etd repository at iitk library using dspace,” in international conference on semantic web and digital libraries (icsd- ), ed. a. r. d. prasad and devika p. madalli ( ), – . http://hdl.handle .net/ / (accessed dec. , ). . todd m. mundle, “digital retrospective conversion of theses and dissertations: an in house project” (paper presented to the th international symposium on electronic theses & dis- sertations, sydney, australia, sept. – , ), http://adt.caul .edu.au/etd /papers/ mundle.pdf (accessed dec. , ). . rowan brownlee, “research data and repository meta- data: policy and technical issues at the university of sydney library,” cataloging & classification quarterly , no. / ( ): – . . steve thomas, “importing marc data into dspace,” , http://hdl.handle.net/ / (accessed dec. , ). . sarah kim, lorraine a. dong, and megan durden, “auto- mated batch archival processing: preserving arnold wesker’s digital manuscripts,” archival issues , no. ( ): – . . elspeth healey, samantha mueller, and sarah ticer, “the paul n. banks papers: archiving the electronic records of a digitally-adventurous conservator,” , https://pacer .ischool.utexas.edu/bitstream/ / / /paul_banks_ final_report.pdf (accessed dec. , ); lisa schmidt, “pres- ervation of a born digital literary genre: archiving legacy macintosh hypertext files in dspace,” , https://pacer .ischool.utexas.edu/bitstream/ / / /mj% wbo% capstone% report.pdf (accessed dec. , ). . rachel e. proudfoot et al., “jisc final report: increase (increasing repository content through automation and ser- vices),” , http://eprints.whiterose.ac.uk/ / (accessed dec. , ). . michael witt and mark p. newton, “preparing batch deposits for digital commons repositories,” , http://docs .lib.purdue.edu/lib_research/ / (accessed dec. , ). . lesley drysdale, “importing records from reference man- ager into gnu eprints,” , http://hdl.handle.net/ / (accessed dec. , ). . r. john robertson, “evaluation of metadata workflows for the glasgow eprints and dspace services,” , http://hdl .handle.net/ / (accessed dec. , ); william j. nixon and morag greig, “populating the glasgow eprints service: a mediated model and workflow,” , http://hdl.handle .net/ / (accessed dec. , ). . tim ribaric, “automatic preparation of etd material from the internet archive for the dspace repository platform,” code lib journal no. (nov. , ), http://journal.code lib.org/ articles/ (accessed dec. , ). . randall floyd, “automated electronic thesis and disser- tations ingest,” (mar. , ), http://wiki.dlib.indiana.edu/ confluence/x/ y (accessed dec. , ). . shawn averkamp and joanna lee, “repurposing pro- batch loading workflows are dependent upon the quality of data and metadata loaded. along with testing scripts and checking imported metadata by first batch loading to a development or staging environment, quality control of the supplied metadata is an integral step. the flexibility of perl allowed testing and revising to accommodate prob- lems encountered with how the metadata was supplied for the heterogeneous collections batch loaded into the knowledge bank. however, toward the goal of standard- izing batch loading workflows, the staff working with the knowledge bank iteratively refined not only the scripts but also the metadata requirements for each project and how those were communicated to the data suppliers with mappings, explicit metadata examples, and sample desired results. the efficiency of batch loading workflows is greatly enhanced by consistent data and basic stan- dards for how metadata is supplied. batch loading is not only an extremely efficient means of populating an institutional repository, it is also a value- added service that can increase buy-in from the wider campus community. it is hoped that by openly sharing examples of our batch loading scripts we are contributing to the development of an open library of code that can be borrowed and adapted by the library community toward future institutional repository success stories. ■■ acknowledgments i would like to thank conrad gratz, of osu oit, and andrew wang, formerly of osul. gratz wrote the shell scripts and the majority of the perl scripts used for auto- mating the knowledge bank item import process and ran the corresponding batch loads. the early perl scripts used for batch loading into the knowledge bank, including the first phase of ojs and mss, were written by wang. parts of those early perl scripts written by wang were borrowed for subsequent scripts written by gratz. gratz provided the annotated scripts appearing in the appendixes and consulted with the author regarding the description of the scripts. i would also like to thank amanda j. wilson, a for- mer metadata librarian for osul, who was instrumental to the success of many of the batch loading workflows created for the knowledge bank. references and notes . the ohio state university knowledge bank, “institu- tional repository policies,” , http://library.osu.edu/sites/ kbinfo/policies.html (accessed dec. , ). the knowledge bank homepage can be found at https://kb.osu.edu/dspace/ (accessed dec. , ). . margret branschofsky et al., “evolving meta- data needs for an institutional repository: mit’s dspace,” information technology and libraries | september appendix e. mss batch loading scripts -- mkxml .pl -- #!/usr/bin/perl use encode; # routines for utf encoding use text::xsv; # routines to process csv files. use file::basename; # open and read the comma separated metadata file. my $csv = new text::xsv; #$csv->set_sep(' '); # use for tab separated files. $csv->open_file("mss .csv"); $csv->read_header(); # process the csv column headers. # constants for file and directory names. $basedir = "/common/batch/input/mss/"; $indir = "$basedir/ "; $xmldir= "./ xml"; $imagesubdir= "processed_images”; $filename = "dublin_core.xml"; # process each line of metadata, one line per item. $linenum = ; while ($csv->get_row()) { # this divides the item's metadata into fields, each in its own variable. my ( $identifier, $title, $creators, $description_abstract, $issuedate, $description, $description , appendixes a–d available at http://ital-ica.blogspot.com/ quest metadata for batch ingesting etds into an institutional repository,” code lib journal no. (june , ), http://journal .code lib.org/articles/ (accessed dec. , ). . tim brody, registry of open access repositories (roar), http://roar.eprints.org/ (accessed dec. , ). . duraspace, dspace, http://www.dspace.org/ (accessed dec. , ). . dublin core metadata initiative libraries working group, “dc-library application profile (dc-lib),” http://dublincore .org/documents/ / / /library-application-profile/ (accessed dec. , ). . the ohio state university knowledge bank policy com- mittee, “osu knowledge bank metadata application profile,” http://library.osu.edu/sites/techservices/kbappprofile.php (accessed dec. , ). . ohio journal of science (ohio academy of sci- ence), knowledge bank community, http://hdl.handle .net/ / (accessed dec. , ); osu international sym- posium on molecular spectroscopy, knowledge bank commu- nity, http://hdl.handle.net/ / (accessed dec. , ). . ohio journal of science (ohio academy of science), ohio journal of science: volume , issue (may, ), knowledge bank collection, http://hdl.handle.net/ / (accessed dec. , ). batch loading collections into dspace | walsh $abstract, $gif, $ppt, ) = $csv->extract( "talk_id", "title", "creators", "abstract", "issuedate", "description", "authorinstitution", "image_file_name", "talk_gifs_file", "talk_ppt_file" ); $creatorxml = ""; # multiple creators are separated by ';' in the metadata. if (length($creators) > ) { # create xml for each creator. @creatorlist = split(/;/,$creators); foreach $creator (@creatorlist) { if (length($creator) > ) { $creatorxml .= '' .$creator.’’.”\n “; } } } # done processing creators for this item. # create the xml string for the abstract. $abstractxml = ""; if (length($description_abstract) > ) { # convert special metadata characters for use in xml/html. $description_abstract =~ s/\&/&/g; $description_abstract =~ s/\>/>/g; $description_abstract =~ s/\' .$description_abstract.''; } # create the xml string for the description. $descriptionxml = ""; if (length($description) > ) { # convert special metadata characters for use in xml/html. $description=~ s/\&/&/g; $description=~ s/\>/>/g; $description=~ s/\' .$description.''; } appendix e. mss batch loading scripts (cont.) information technology and libraries | september # create the xml string for the author institution. $description xml = ""; if (length($description ) > ) { # convert special metadata characters for use in xml/html. $description =~ s/\&/&/g; $description =~ s/\>/>/g; $description =~ s/\' .'author institution: '.$description .''; } # convert special characters in title. $title=~ s/\&/&/g; $title=~ s/\>/>/g; $title=~ s/\:encoding(utf- )", "$basedir/$subdir/$filename"); print fh <<"xml"; $identifier $title $issuedate $abstractxml $descriptionxml $description xml article en $creatorxml xml close($fh); # create contents file and move files to the load set. # copy item files into the load set. if (defined($abstract) && length($abstract) > ) { system "cp $indir/$abstract $basedir/$subdir"; } $sourcedir = substr($abstract, , ); if (defined($ppt) && length($ppt) > ) { system "cp $indir/$sourcedir/$sourcedir/*.* $basedir/$subdir/"; } if (defined($gif) && length($gif) > ) { system "cp $indir/$sourcedir/$imagesubdir/*.* $basedir/$subdir/"; } # make the 'contents' file and fill it with the file names. appendix e. mss batch loading scripts (cont.) batch loading collections into dspace | walsh system "touch $basedir/$subdir/contents"; if (defined($gif) && length($gif) > && -d "$indir/$sourcedir/$imagesubdir" ) { # sort items in reverse order so they show up right in dspace. # this is a hack that depends on how the db returns items # in unsorted (physical) order. there are better ways to do this. system "cd $indir/$sourcedir/$imagesubdir/;" . " ls *[ - ][ - ].* | sort -r >> $basedir/$subdir/contents"; system "cd $indir/$sourcedir/$imagesubdir/;" . " ls *[a-za-z][ - ].* | sort -r >> $basedir/$subdir/contents"; } if (defined($ppt) && length($ppt) > && -d "$indir/$sourcedir/$sourcedir" ) { system "cd $indir/$sourcedir/$sourcedir/;" . " ls *.* >> $basedir/$subdir/contents"; } # put the abstract in last, so it displays first. system "cd $basedir/$subdir; basename $abstract >>" . " $basedir/$subdir/contents"; $linenum++; } # done processing an item. -------------------------------------------------------------------------------------------------- -- import.sh –- #!/bin/sh # # import a collection from files generated on dspace # collection_id= / eperson=[name removed]@osu.edu source_dir=./ xml base_id=`basename $collection_id` mapfile=./map-dspace -mss .$base_id /dspace/bin/dsrun org.dspace.app.itemimport.itemimport --add --eperson=$eperson --collection=$collection_id --source=$source_dir --mapfile=$mapfile appendix e. mss batch loading scripts (cont.) information technology and libraries | september lynne weber and peg lawrence authentication and access: accommodating public users in an academic world in cook and shelton’s managing public computing, which confirmed the lack of applicable guidelines on academic websites, had more up-to-date information but was not available to the researchers at the time the project was initiated. in the course of research, the authors developed the following questions: ■■ how many arl libraries require affiliated users to log into public computer workstations within the library? ■■ how many arl libraries provide the means to authenticate guest users and allow them to log on to the same computers used by affiliates? ■■ how many arl libraries offer open-access comput- ers for guests to use? do these libraries provide both open-access computers and the means for guest user authentication? ■■ how do federal depository library program libraries balance their policy requiring computer authentica- tion with the obligation to provide public access to government information? ■■ do computers provided for guest use (open access or guest login) provide different software or capabilities than those provided to affiliated users? ■■ how many arl libraries have written policies for the use of open-access computers? if a policy exists, what is it? ■■ how many arl libraries have written policies for authenticating guest users? if a policy exists, what is it? ■■ literature review since the s there has been considerable discussion within library literature about academic libraries serving “external,” “secondary,” or “outside” users. the subject has been approached from the viewpoint of access to the library facility and collections, reference assistance, interlibrary loan (ill) service, borrowing privileges, and (more recently) access to computers and internet privi- leges, including the use of proprietary databases. deale emphasized the importance of public relations to the academic library. while he touched on creating bonds both on and off campus, he described the positive effect of “privilege cards” to community members. josey described the variety of services that savannah state college offered to the community. he concluded his essay with these words: why cannot these tried methods of lending books to citizens of the community, story hours for children . . . , a library lecture series or other forum, a great books discussion group and the use of the library staff in the fall of , the academic computing center, a division of the information technology services department (its) at minnesota state university, mankato took over responsibility for the computers in the public areas of memorial library. for the first time, affiliated memorial library users were required to authenticate using a campus username and password, a change that effectively eliminated computer access for anyone not part of the university community. this posed a dilemma for the librarians. because of its federal depository status, the library had a responsibility to pro- vide general access to both print and online government publications for the general public. furthermore, the library had a long tradition of providing guest access to most library resources, and there was reluctance to aban- don the practice. therefore the librarians worked with its to retain a small group of six computers that did not require authentication and were clearly marked for community use, along with several standup, open-access computers on each floor used primarily for searching the library catalog. the additional need to provide computer access to high school students visiting the library for research and instruction led to more discussions with its and resulted in a means of generating temporary usernames and passwords through a web form. these user accommodations were implemented in the library without creating a written policy governing the use of open-access computers. o ver time, library staff realized that guidelines for guests using the computers were needed because of misuse of the open-access computers. we were charged with the task of drafting these guidelines. in typical librarian fashion, we searched websites, including those of association of research libraries (arl) members for existing computer access policies in academic libraries. we obtained very little information through this search, so we turned to arl publications for assistance. library public access workstation authentication by lori driscoll, was of greater benefit and offered much of the needed information, but it was dated. a research result described lynne webber (lnweber@mnsu.edu) is access services librar- ian and peg lawrence (peg.lawrence@mnsu.edu) is systems librarian, minnesota state university, mankato. authentication and access | weber and lawrence providing service to the unaffiliated, his survey revealed percent of responding libraries offered free in-house collection use for the general public, and many others offered additional services. brenda johnson described a one-day program in sponsored by rutgers university libraries forum titled “a case study in closing the university library to the public.” the participating librarians spent the day famil- iarizing themselves with the “facts” of the theoretical case and concluded that public access should be restricted but not completely eliminated. a few months later, consider- ation of closing rutgers’ library to the public became a real debate. although there were strong opposing view- points, the recommendation was to retain the open-door policy. jansen discussed the division between those who wanted to provide the finest service to primary users and those who viewed the library’s mission as including all who requested assistance. jansen suggested specific ways to balance the needs of affiliates and the public and referred to the dilemma the university of california, berkeley, library that had been closed to unaffiliated users. bobp and richey determined that california undergraduate libraries were emphasizing service to pri- mary users at a time when it was no longer practical to offer the same level of service to primary and secondary users. they presented three courses of action: adherence to the status quo, adoption of a policy restricting access, or implementation of tiered service. throughout the s, the debate over the public’s right to use academic libraries continued, with increasing focus on computer use in public and private academic libraries. new authorization and authentication require- ments increased the control of internal computers, but the question remained of libraries providing access to government information and responding to community members who expected to use the libraries supported by their taxes. morgan, who described himself as one who had spent his career encouraging equal access to information, con- cluded that it would be necessary to use authentication, authorization, and access control to continue offering information services readily available in the past. martin acknowledged that library use was changing as a result of the internet and that the public viewed the academic librarian as one who could deal with the explosion of information and offer service to the public. johnson described unaffiliated users as a group who wanted all the privileges of the affiliates; she discussed the obliga- tion of the institution to develop policies managing these guest users. still and kassabian considered the dual responsi- bilities of the academic library to offer internet access to public users and to control internet material received and sent by primary and public users. further, they weighed as consultants be employed toward the building of good relations between town and gown. later, however, deale indicated that the generosity common in the s to outsiders was becoming unsus- tainable. deale used beloit college, with an “open door policy” extending more than years, as an example of a school that had found it necessary to refuse out-of-library circulation to minors except through ill by the s. also in , waggoner related the increasing difficulty of accommodating public use of the academic library. he encouraged a balance of responsibility to the public with the institution’s foremost obligation to the students and faculty. in october , the ad hoc committee on community use of academic libraries was formed by the college library section of the association of college and research libraries (acrl). this committee distributed a -ques- tion survey to , colleges and universities throughout the united states. the high rate of response ( per- cent) was considered noteworthy, and the findings were explored in “community use of academic libraries: a symposium,” published in . the concluding article by josey (the symposium’s moderator) summarized the lenient attitudes of academic libraries toward public users revealed through survey and symposium reports. in the same article, josey followed up with his own arguments in favor of the public’s right to use academic libraries because of the state and federal support provided to those institutions. similarly, in tolliver reported the results of a survey of wisconsin libraries (public academic, private academic, and public), which indicated that respondents made a great effort to serve all patrons seeking service. tolliver continued in a different vein from josey, however, by reporting the current annual fiscal support for libraries in wisconsin and commenting upon financial steward- ship. tolliver concluded by asking, “how effective are our library systems and cooperative affiliations in meet- ing the information needs of the citizens of wisconsin?” much of the literature in the years following focused on serving unaffiliated users at a time when public and academic libraries suffered the strain of overuse and underfunding. the need for prioritization of primary users was discussed. in , russell asked, “who are our legitimate clientele?” and countered the argument for publicly supported libraries serving the entire public by saying the public “cannot freely use the university lawn mowers, motor pool vehicles, computer center, or athletic facilities.” ten years later, russell, robison, and prather prefaced their report on a survey of policies and services for outside users at consortia institutions by saying, “the issue of external users is of mounting concern to an institution whose income is student credit hour gen- erated.” despite russell’s concerns about the strain of information technology and libraries | september be aware of the issues and of the effects that licensing, networking, and collection development decisions have on access.” in “unaffiliated users’ access to academic libraries: a survey,” courtney reported and analyzed data from her own comprehensive survey sent to academic libraries in winter . of the libraries responding to the survey, libraries ( . percent) required all users to authenticate to use computers within the library, while ( . percent) indicated that they planned to require authentication in the next twelve months. courtney followed this with data from surveyed libraries that had canceled “most” of their indexes and abstracts ( librar- ies, or . percent) and libraries that had cancelled “most” periodicals ( libraries or . percent). she concluded that the extent to which the authentication requirement restricted unaffiliated users was not clear, and she asked, “as greater numbers of resources shift to electronic-only formats, is it desirable that they disappear from the view of the community user or the visiting scholar?” courtney’s “authentication and library public access computers: a call for discussion” described a follow-up with the academic libraries participating in her survey who had self-identified as using authentication or planning to employ authentication within the next twelve months. her conclusion was the existence of ambivalence toward authentication among the libraries, since more than half of the respondents provided some sort of public access. she encouraged librarians to carefully consider the library’s commitment to service before entering into blanket license agreements with vendors or agreeing to campus computer restrictions. several editions of the arl spec kit series showing trends of authentication and authorization for all users of arl libraries have been an invaluable resource in this investigation. an examination of earlier spec kits indicated that the definitions of “user authentication” and “authorization” have changed over the years. user authentication, by plum and bleiler indicated that per- cent of surveyed libraries authenticated users in some way, but at that time authentication would have been more precisely defined as authorization or permission to access personal records, such as circulation, e-mail, course regis- tration, and file space. as such, neither authentication nor authorization was related to basic computer access. by contrast, it is common for current library users authenti- cate to have any access to a public workstation. driscoll’s library public access workstation authentication sought information on how and why users were authenticated on public-access computers, who was driving the change, how it affected the ability of federal depository libraries to provide public information, and how it affected library ser- vices in general. but at the time of driscoll’s survey, only percent of surveyed libraries required authentication on all computers and percent required it only on selected terminals. cook and shelton’s managing public computing the reconciliation of material restrictions against “prin- ciples of freedom of speech, academic freedom, and the ala’s condemnation of censorship.” lynch discussed institutional use of authentication and authorization and the growing difficulty of verifying bona fide users of aca- demic library subscription databases and other electronic resources. he cautioned that future technical design choices must reflect basic library values of free speech, personal confidentiality, and trust between academic institution and publisher. barsun specifically examined the webpages of one hundred arl libraries in search of information pertinent to unaffiliated users. she included a historic overview of the changing attitudes of academics toward service to the unaffiliated population and described the difficult bal- ance of college community needs with those of outsiders in (the survey year). barsun observed a consistent lack of information on library websites regarding library guest use of proprietary databases. carlson discussed academic librarians’ concerns about “internet-related crimes and hacking” leading to reconsideration of open computer use, and he described the need to compromise patron privacy by requiring authentication. in a chapter on the relationship of it security to academic values, oblinger said, “one possible interpretation of intellectual freedom is that individuals have the right to open and unfiltered access to the internet.” this statement was followed later with “equal access to information can also be seen as a logical extension of fairness.” a short article in library and information update alerted the authors to a uk project investigating improved online access to resources for library visitors not affili- ated with the host institution. salotti described higher education access to e-resources in visited institutions (haervi) and its development of a toolkit to assist with the complexities of offering electronic resources to guest users. salotti summarized existing resources for sharing within the united kingdom and emphasized that “no single solution is likely to suit all universities and col- leges, so we hope that the toolkit will offer a number of options.” launched by the society of college, national and university libraries (sconul), and universities and colleges information systems association (ucisa), haervi has created a best-practice guide. by far the most useful articles for this investigation have been those by nancy courtney. “barbarians at the gates: a half-century of unaffiliated users in academic libraries,” a literature review on the topic of visitors in academic libraries, included a summary of trends in attitude and practice toward visiting users since the s. the article concluded with a warning: “the shift from printed to elec- tronic formats . . . combined with the integration of library resources with campus computer networks and the internet poses a distinct threat to the public’s access to information even onsite. it is incumbent upon academic librarians to authentication and access | weber and lawrence introductory letter with the invitation to participate and a forward containing definitions of terms used within the survey is in appendix a. in total, ( percent) of the arl libraries invited to participate in the survey responded. this is comparable with the response rate for similar surveys reported by plum and bleiler ( of , or percent), driscoll ( of , or percent), and cook and shelton ( of , or percent). . what is the name of your academic institution? the names of the responding libraries are listed in appendix b. . is your institution public or private? see figure . respondents’ explanations of “other” are listed below. ■❏ state-related ■❏ trust instrument of the u.s. people; quasi- government ■❏ private state-aided ■❏ federal government research library ■❏ both—private foundation, public support . are affiliated users required to authenticate in order to access computers in the public area of your library? see figure . . if you answered “yes” to the previous question, does your library provide the means for guest users to authenticate? see figure . respondents’ explanations of “other” are listed below. all described open-access comput- ers. ■❏ “we have a few “open” terminals” ■❏ “ computers don’t require authentication” ■❏ “some workstations do not require authentica- tion” ■❏ “open-access pcs for guests (limited number and function)” ■❏ “no—but we maintain several open pcs for guests” ■❏ “some workstations do not require login” . is your library a federal depository library? see fig- ure . this question caused some confusion for the canadian survey respondents because canada has its own depository services program corresponding to the u.s. federal depository program. consequently, of the respondents identified themselves as federal depository (including three canadian librar- ies), although of the are more accurately mem- bers of the canadian depository services program. only two responding libraries were neither a mem- ber of the u.s. federal depository program nor of the canadian depository services program. . if you answered “yes” to the previous question, and com- puter authentication is required, what provisions have been made to accommodate use of online government documents by the general public in the library? please check all that touched on every aspect of managing public computing, including public computer use, policy, and security. even in , only percent of surveyed libraries required authentication on all computers, but percent required authentication on some computers, showing the trend toward an ever increasing number of libraries requiring public workstation authentication. most of the responding libraries had a computer-use policy, with percent follow- ing an institution-wide policy developed by the university or central it department. ■■ method we constructed a survey designed to obtain current data about authentication in arl libraries and to provide insight into how guest access is granted at various aca- demic institutions. it should be noted that the object of the survey was access to computers located in the public areas of the library for use by patrons, not access to staff computers. we constructed a simple, fourteen-question survey using the zoomerang online tool (http://www .zoomerang.com/). a list of the deans, directors, and chief operating officers from the arl libraries was compiled from an internet search. we eliminated the few library administrators whose addresses could not be readily found and sent the survey to individuals with the request that it be forwarded to the appropriate respondent. the recipients were informed that the goal of the project was “determination of computer authentica- tion and current computer access practices within arl libraries” and that the intention was “to reflect practices at the main or central library” on the respondent’s cam- pus. recipients were further informed that the names of the participating libraries and the responses would be reported in the findings, but that there would be no link between responses given and the name of the participat- ing library. the survey introduction included the name and contact information of the institutional review board administrator for minnesota state university, mankato. potential respondents were advised that the e-mail served as informed consent for the study. the survey was administered over approximately three weeks. we sent reminders three, five, and seven days after the survey was launched to those who had not already responded. ■■ survey questions, responses, and findings we administered the survey, titled “authentication and access: academic computers . ,” in late april . following is a copy of the fourteen-question survey with responses, interpretative data, and comments. the information technology and libraries | september ■❏ “some computers are open access and require no authentication” ■❏ “some workstations do not require login” . if your library has open-access computers, how many do you provide? (supply number). see figure . a total of institutions responded to this question, and reported open-access computers. the number of open-access computers ranged from to , . as expected, the highest numbers were reported by libraries that did not require authentication for affili- ates. the mean number of open-access computers was . , the median was , the mode was , and the range was , . . please indicate which online resources and services are available to authenticated users. please check all that apply. see figure . ■❏ online catalog ■❏ government documents ■❏ internet browser apply. see figure . ■❏ temporary user id and password ■❏ open access computers (unlimited access) ■❏ open access computers (access limited to government documents) ■❏ other of the libraries that responded “yes” to question , required authentication for affiliates. these institutions offered the general public access to online government documents various ways. explanations of “other” are listed below. three of these responses indicate, by survey definition, that open-access computers were provided. ■❏ “catalog-only workstations” ■❏ “ computers don’t require authentication” ■❏ “generic login and password” ■❏ “librarians login each guest individually” ■❏ “provision made for under- guests needing gov doc” ■❏ “staff in gov info also login user for quick use” ■❏ “restricted guest access on all public devices” figure . institutions with the means to authenticate guests figure . libraries with federal depository and/or canadian depository services status figure . institutions requiring authentication figure . categories of responding institutions authentication and access | weber and lawrence . does your library have a written policy for use of open access computers in the public area of the library? question indicates that of the responding libraries did offer the public two or more open-access computers. out of the , responded that they had a written policy governing the use of computers. conversely, open-access computers were reported at libraries that had no reported written policy. . if you answered “yes” to the previous question, please give the link to the policy and/or summarize the policy. twenty-eight libraries gave a url, a url plus a summary explanation, or a summary explanation with no url. . does your library have a written policy for authenticating guest users? out of the libraries that required their users to authenticate (see question ), also had the means to allow their guests to authenticate (see question ). fifteen of those libraries said they had a policy. . if you answered “yes” to the previous question, please give the link to the policy and/or summarize the policy. eleven ■❏ licensed electronic resources ■❏ personal e-mail access ■❏ microsoft office software . please indicate which online resources and services are available to authenticated guest users. please check all that apply. see figure . ■❏ online catalog ■❏ government documents ■❏ internet browser ■❏ licensed electronic resources ■❏ personal e-mail access ■❏ microsoft office software . please indicate which online resources and services are available on open-access computers. please check all that apply. see figure . ■❏ online catalog ■❏ government documents ■❏ internet browser ■❏ licensed electronic resources ■❏ personal e-mail access ■❏ microsoft office software figure . provisions for the online use of government docu- ments where authentication is required figure . number of open-access computers offered figure . electronic resources for authenticated affiliated users (n = ) number of libraries number of librariesnumber of libraries number of libraries figure . resources for authenticating guest users (n = ) information technology and libraries | september ■■ respondents and authentication figure compares authentication practices of public, private, and other institutions described in response to question . responses from public institutions outnum- bered those from private institutions, but within each group a similar percentage of libraries required their affiliated users to authenticate. therefore no statistically significant difference was found between authenticating affiliates in public and private institutions. of the respondents, ( percent) required their affiliated users to authenticate (see question ) and of the also had the means to authenticate guests (see question ). the remaining offered open-access comput- ers. fourteen libraries had both the means to authenticate guests and had open-access computers (see questions and ). when we compare the results of the study by cook and shelton with the results of the current study (completed in ), the results are somewhat contradic- tory (see table ). the differences in survey data seem to indicate that authentication requirements are decreasing; however, the literature review—specifically cook and shelton and the courtney article—clearly indicate that authentica- tion is on the rise. this dichotomy may be explained, in part, by the fact that of the more than arl libraries responding to both surveys, there was an overlap of only libraries. the u.s. federal depository or canadian depository services libraries that required their affiliated users to authenticate (see questions and ) provided guest access ranging from usernames and passwords, to open-access computers, to computers restricted to libraries gave the url to their policy; summarized their policies. ■■ research questions answered the study resulted in answers to the questions we posed at the outset: ■■ thirty-two ( percent) of the responding arl libraries required affiliated users to login to public computer workstations in the library. ■■ twenty-three ( percent) of the arl libraries requiring affiliated users to login to public computers provided the means for guest users to login to public computer workstations in the library. ■■ fifty ( percent) of responding arl libraries provided open-access computers for guest users; ( percent) of those libraries provided both open-access computers and the means for guest authentication. ■■ without exception, all u.s. federal depository or canadian depository services libraries that required their users to authenticate offered guest users some form of access to online information. ■■ survey results indicated some differences between software provided to various users on differently accessed computers. office software was less fre- quently provided on open-access computers. ■■ twenty-eight responding arl libraries had written policies relating to the use of open-access computers. ■■ fifteen responding arl libraries had written policies relating to the authorization of guests. figure . electronic resources on open access computers (n = ) figure . comparison of library type and authentication requirement number of libraries authentication and access | weber and lawrence ■■ one library had guidelines for use posted next to the workstations but did not give specifics. ■■ fourteen of those requiring their users to authen- ticate had both open-access computers and guest authentication to offer to visitors of their libraries. other policy information was obtained by an exami- nation of the websites listed by respondents: ■■ ten of the sites specifically stated that the open-access computers were for academic use only. ■■ five of the sites specified time limits for use of open- access computers, ranging from to minutes. ■■ four stated that time limits would be enforced when others were waiting to use computers. ■■ one library used a sign-in sheet to monitor time limits. ■■ one library mentioned a reservation system to moni- tor time limits. ■■ two libraries prohibited online gambling. ■■ six libraries prohibited viewing sexually explicit materials. ■■ guest-authentication policies of the libraries that had the means to authenticate their guests, had a policy for guests obtaining a username and password to authenticate, and outlined their requirements of showing identification and issuing access. the other had open-access computers that guests might use. the following are some of the varied approaches to guest authentication: ■■ duration of the access (when mentioned) ranged from days to months. ■■ one library had a form of sponsored access where current faculty or staff could grant a temporary user- name and password to a visitor. ■■ one library had an online vouching system that allowed the visitor to issue his or her own username and password online. ■■ one library allowed guests to register themselves by swiping an id or credit card. ■■ one library had open-access computers for local resources and only required authentication to leave the library domain. ■■ one library had the librarians log the users in as guests. ■■ one library described the privacy protection of col- lected personal information. ■■ no library mentioned charging a fee for allowing computer access. government documents, to librarians logging in for guests (see question ). numbers of open-access comput- ers ranged widely from to more than , (see question ). eleven ( percent) of the responding u.s. federal depository or canadian depository services libraries that did not provide open-access computers issued a tempo- rary id (nine libraries), provided open access limited to government documents (one library), or required librar- ian login for each guest (one library). all libraries with u.s. federal depository or canadian depository services status provided a means of public access to information to fulfill their obligation to offer government documents to guests. figure shows a comparison of resources available to authenticated users and authenticated guests and offered on open-access computers. as might be expected, almost all institutions provided access to online catalogs, government documents, and internet browsers. fewer allowed access to licensed electronic resources and e-mail. access to office software showed the most dramatic drop in availability, especially on open-access computers. ■■ open-access computer policies as mentioned earlier, libraries had written policies for their open-access computers (see question ), and libraries gave a url, a url plus a summary explanation, or a summary explanation with no url (see question ). in most instances, the library policy included their campus’s acceptable-use policy. seven libraries cited their campus’s acceptable-use policy and nothing else. nearly all libraries applied the same acceptable-use policy to all users on all computers and made no distinction between policies for use of open-access computers or computers requiring authentication. following are some of the varied aspects of summa- rized policies pertaining to open-access computers: ■■ eight libraries stated that the computers were for aca- demic use and that users might be asked to give up their workstation if others were waiting. table . comparison of findings from cook and shelton ( ) and the current survey ( ) authentication requirements (n = ) (n = ) some required ( %) ( %) required for all ( %) ( %) not required ( %) ( %) information technology and libraries | september ■■ further study although the survey answered many of our questions, other questions arose. while the number of libraries requiring affiliated users to log on to their public com- puters is increasing, this study does not explain why this is the case. reasons could include reactions to the september disaster, the usa patriot act, general security concerns, or the convenience of the personalized desktop and services for each authenticated user. perhaps a future investigation could focus on reasons for more frequent requirement of authentication. other subjects that arose in the examination of institutional policies were guest fees for services, age limits for younger users, com- puter time limits for guests, and collaboration between academic and public libraries. ■■ policy developed as a result of the survey findings as a result of what was learned in the survey, we drafted guidelines governing the use of open-access computers by visitors and other non-university users. the guidelines can be found at http://lib.mnsu.edu/about/libvisitors .html#access. these guidelines inform guests that open- access computers are available to support their research, study, and professional activities. the computers also are governed by the campus policy and the state university system acceptable-use policy. guideline provisions enable staff to ask users to relinquish a computer when others are waiting or if the computer is not being used for academic purposes. while this library has the ability to generate temporary usernames and passwords, and does so for local schools coming to the library for research, no guide- lines have yet been put in place for this function. figure . online resources available to authenticated affiliated users, guest users, open-access users authentication and access | weber and lawrence these practices depend on institutional missions and goals and are limited by reasonable considerations. in the past, accommodation at some level was generally offered to the community, but the complications of affili- ate authentication, guest registration, and vendor-license restrictions may effectively discourage or prevent outside users from accessing principal resources. on the other hand, open-access computers facilitate access to electronic resources. those librarians who wish to provide the same level of commitment to guest users as in the past as well as protect the rights of all should advocate to campus policy-makers at every level to allow appropriate guest access to computers to fulfill the library’s mission. in this way, the needs and rights of guest users can be balanced with the responsibilities of using campus computers. in addition, librarians should consider ensuring that the licenses of all electronic resources accommodate walk-in users and developing guidelines to prevent incor- poration of electronic materials that restrict such use. this is essential if the library tradition of freedom of access to information is to continue. finally, in regard to external or guest users, academic librarians are pulled in two directions; they are torn between serving primary users and fulfilling the prin- ciples of intellectual freedom and free, universal access to information along with their obligations as federal depository libraries. at the same time, academic librar- ians frequently struggle with the goals of the campus administration responsible for providing secure, reliable networks, sometimes at the expense of the needs of the outside community. the data gathered in this study, indicating that percent of responding libraries con- tinue to provide at least some open-access computers, is encouraging news for guest users. balancing public access and privacy with institutional security, while a current concern, may be resolved in the way of so many earlier preoccupations of the electronic age. given the pervasiveness of the problem, however, fair and equitable treatment of all library users may continue to be a central concern for academic libraries for years to come. references . lori driscoll, library public access workstation authentica- tion, spec kit (washington, d.c.: association of research libraries, ). . martin cook and mark shelton, managing public comput- ing, spec kit (washington, d.c.: association of research libraries, ): . . h. vail deale, “public relations of academic libraries,” library trends (oct. ): – . . ibid., . . e. j. josey, “the college library and the community,” faculty research edition, savannah state college bulletin (dec. ): – . ■■ conclusions while we were able to gather more than years of litera- ture pertaining to unaffiliated users in academic libraries, it soon became apparent that the scope of consideration changed radically through the years. in the early years, there was discussion about the obligation to provide service and access for the community balanced with the challenge to serve two clienteles. despite lengthy debate, there was little exception to offering the community some level of service within academic libraries. early preoccupation with physical access, material loans, ill, basic reference, and other services later became a discus- sion of the right to use computers, electronic resources, and other services without imposing undue difficulty to the guest. current discussions related to guest users reflect obvious changes in public computer administration over the years. authentication presently is used at a more fundamental level than in earlier years. in many librar- ies, users must be authorized to use the computer in any way whatsoever. as more and more institutions require authentication for their primary users, accommodation must be made if guests are to continue being served. in addition, as courtney’s research indicates, an ever increasing number of electronic databases, indexes, and journals replace print resources in library collections. this multiplies the roadblocks for guest users and exacerbates the issue. unless special provisions are made for com- puter access, community users are left without access to a major part of the library’s collections. because of the arl libraries ( percent) are federal depository or canadian depository services libraries, the researchers hypothesized that most librar- ies responding to the survey would offer open-access computers for the use of nonaffiliated patrons. this study has shown that federal depository libraries have remained true to their mission and obligation of provid- ing public access to government-generated documents. every federal depository respondent indicated that some means was in place to continue providing visitor and guest access to the majority of their electronic resources— whether through open-access computers, temporary or guest logins, or even librarians logging on for users. while access to government resources is required for the librar- ies housing government-document collections, libraries can use considerably more discretion when considering what other resources guest patrons may use. despite the commitment of libraries to the dissemination of govern- ment documents, the increasing use of authentication may ultimately diminish the libraries’ ability and desire to accommodate the information needs of the public. this survey has provided insight into the various ways academic libraries serve guest users. not all academic libraries provide public access to all library resources. information technology and libraries | september identify yourself,” chronicle of higher education , no. (june , ): a , http://search.ebscohost.com/login.aspx?direct =true&db=aph&an= &site=ehost-live (accessed mar. , ). . diana oblinger, “it security and academic values,” in luker and petersen, computer & network security in higher edu- cation, , http://net.educause.edu/ir/library/pdf/pub e .pdf (accessed july , ). . ibid., . . “access for non-affiliated users,” library & information update , no. ( ): . . paul salotti, “introduction to haervi-he access to e-resources in visited institutions,” sconul focus no. (dec. ): – , http://www.sconul.ac.uk/publications/ newsletter/ / .pdf (accessed july , ). . ibid., . . universities and colleges information systems asso- ciation (ucisa), haervi: he access to e-resources in visited institutions, (oxford: ucisa, ), http://www.ucisa.ac.uk/ publications/~/media/files/members/activities/haervi/ haerviguide% pdf (accessed july , ). . nancy courtney, “barbarians at the gates: a half-century of unaffiliated users in academic libraries,” journal of academic librarianship , no. (nov. ): – , http://search.ebsco host.com/login.aspx?direct=true&db=aph&an= &site= ehost-live (accessed july , ). . ibid., . . nancy courtney, “unaffiliated users’ access to academic libraries: a survey,” journal of academic librarianship , no. (jan. ): – , http://search.ebscohost.com/login.aspx?dire ct=true&db=aph&an= &site=ehost-live (accessed july , ). . ibid., . . ibid., . . ibid., . . nancy courtney, “authentication and library public access computers: a call for discussion,” college & research libraries news , no. (may ): – , , www.ala .org/ala/mgrps/divs/acrl/publications/crlnews/ /may/ authentication.cfm (accessed july , ). . terry plum and richard bleiler, user authentication, spec kit (washington, d.c.: association of research libraries, ): . . lori driscoll, library public access workstation authentica- tion, spec kit (washington, d.c.: association of research libraries, ): . . cook and shelton, managing public computing. . ibid., . . plum and bleiler, user authentication, ; driscoll, library public access workstation authentication, ; cook and shelton, managing public computing, . . cook and shelton, managing public computing, . . ibid.; courtney, unaffiliated users, – . . courtney, unaffiliated users, – . . ibid., . . h. vail deale, “campus vs. community,” library journal (apr. , ): – . . ibid., . . john waggoner, “the role of the private university library,” north carolina libraries (winter ): – . . e. j. josey, “community use of academic libraries: a symposium,” college & research libraries , no. (may ): – . . e. j. josey, “implications for college libraries,” in “com- munity use of academic libraries,” – . . don l. tolliver, “citizens may use any tax-supported library?” wisconsin library bulletin (nov./dec. ): . . ibid., . . ralph e. russell, “services for whom: a search for iden- tity,” tennessee librarian: quarterly journal of the tennessee library association , no. (fall ): , . . ralph e. russell, carolyn l. robison, and james e. prather, “external user access to academic libraries,” the southeastern librarian (winter ): . . ibid., . . brenda l. johnson, “a case study in closing the univer- sity library to the public,” college & research library news , no. (sept. ): – . . lloyd m. jansen, “welcome or not, here they come: unaffiliated users of academic libraries,” reference services review , no. (spring ): – . . mary ellen bobp and debora richey, “serving secondary users: can it continue?” college & undergraduate libraries , no. ( ): – . . eric lease morgan, “access control in libraries,” com- puters in libraries , no. (mar. , ): – , http://search .ebscohost.com/login.aspx?direct=true&db=aph&an= & site=ehost-live (accessed aug. , ). . susan k. martin, “a new kind of audience,” journal of academic librarianship , no. (nov. ): , library, infor- mation science & technology abstracts, http://search.ebsco host.com/login.aspx?direct=true&db=aph&an= &site= ehost-live (accessed aug. , ). . peggy johnson, “serving unaffiliated users in publicly funded academic libraries,” technicalities , no. (jan. ): – . . julie still and vibiana kassabian, “the mole’s dilemma: ethical aspects of public internet access in academic libraries,” internet reference services quarterly , no. ( ): . . clifford lynch, “authentication and trust in a networked world,” educom review , no. (jul./aug. ), http://search .ebscohost.com/login.aspx?direct=true&db=aph&an= &site=ehost-live (accessed july , ). . rita barsun, “library web pages and policies toward ‘outsiders’: is the information there?” public services quarterly , no. ( ): – . . ibid., . . scott carlson, “to use that library computer, please authentication and access | weber and lawrence appendix a. the survey introduction, invitation to participate, and forward dear arl member library, as part of a professional research project, we are attempting to determine computer authentication and current com- puter access practices within arl libraries. we have developed a very brief survey to obtain this information which we ask one representative from your institution to complete before april , . the survey is intended to reflect practices at the main or central library on your campus. names of libraries responding to the survey may be listed but no identifying information will be linked to your responses in the analysis or publication of results. if you have any questions about your rights as a research participant, please contact anne blackhurst, minnesota state university, mankato irb administrator. anne blackhurst, irb administrator minnesota state university, mankato college of graduate studies & research alumni foundation mankato, mn ( ) - anne.blackhurst@mnsu.edu you may preview the survey by scrolling to the text below this message. if, after previewing you believe it should be handled by another member of your library team, please forward this message appropriately. alternatively, you may print the survey, answer it manually and mail it to: systems/ access services survey library services minnesota state university, mankato ml —po box mankato, mn - (usa) we ask you or your representative to take minutes to answer questions about computer authentication practices in your main library. participation is voluntary, but follow-up reminders will be sent. this e-mail serves as your informed consent for this study. your participation in this study includes the completion of an online survey. your name and iden- tity will not be linked in any way to the research reports. clicking the link to take the survey shows that you understand you are participating in the project and you give consent to our group to use the information you provide. you have the right to refuse to complete the survey and can discontinue it at any time. to take part in the survey, please click the link at the bottom of this e-mail. thank you in advance for your contribution to our project. if you have questions, please direct your inquiries to the contacts given below. thank you for responding to our invitation to participate in the survey. this survey is intended to determine current academic library practices for computer authentication and open access. your participation is greatly appreciated. below are the definitions of terms used within this survey: ■■ “authentication”: a username and password are required to verify the identity and status of the user in order to log on to computer workstations in the library. ■■ “affiliated user”: a library user who is eligible for campus privileges. ■■ “non-affiliated user”: a library user who is not a member of the institutional community (an alumnus may be a non- affiliated user). this may be used interchangeably with “guest user.” ■■ “guest user”: visitor, walk-in user, nonaffiliated user. ■■ “open access computer”: computer workstation that does not require authentication by user. information technology and libraries | september appendix b. responding institutions . university at albany state university of new york . university of alabama . university of alberta . university of arizona . arizona state university . boston college . university of british columbia . university at buffalo, state university of ny . case western reserve university . university of california berkeley . university of california, davis . university of california, irvine . university of chicago . university of colorado at boulder . university of connecticut . columbia university . dartmouth college . university of delaware . university of florida . florida state university . university of georgia . georgia tech . university of guelph . howard university . university of illinois at urbana-champaign . indiana university bloomington . iowa state university . johns hopkins university . university of kansas . university of louisville . louisiana state university . mcgill university . university of maryland . university of massachusetts amherst . university of michigan . michigan state university . university of minnesota . university of missouri . massachusetts institute of technology . national agricultural library . university of nebraska-lincoln . new york public library . northwestern university . ohio state university . oklahoma state university . university of oregon . university of pennsylvania . university of pittsburgh . purdue university . rice university . smithsonian institution . university of southern california . southern illinois university carbondale . syracuse university . temple university . university of tennessee . texas a&m university . texas tech university . tulane university . university of toronto . vanderbilt university the next generation library catalog | yang and hofmann sharon q. yang and melissa a. hofmann the next generation library catalog: a comparative study of the opacs of koha, evergreen, and voyager open source has been the center of attention in the library world for the past several years. koha and evergreen are the two major open-source integrated library sys- tems (ilss), and they continue to grow in maturity and popularity. the question remains as to how much we have achieved in open-source development toward the next-generation catalog compared to commercial systems. little has been written in the library literature to answer this question. this paper intends to answer this question by comparing the next-generation features of the opacs of two open-source ilss (koha and evergreen) and one proprietary ils (voyager’s webvoyage). m uch discussion has occurred lately on the next- generation library catalog, sometimes referred to as the library . catalog or “the third generation catalog.” different and even conflicting expectations exist as to what the next-generation library catalog comprises: in two sentences, this catalog is not really a catalog at all but more like a tool designed to make it easier for students to learn, teachers to instruct, and scholars to do research. it provides its intended audience with a more effective means for finding and using data and information. such expectations, despite their vagueness, eventually took concrete form in . among the most prominent features of the next-generation catalog are a simple keyword search box, enhanced browsing possibilities, spelling corrections, relevance ranking, faceted naviga- tion, federated search, user contribution, and enriched content, just to mention a few. over the past three years, libraries, vendors, and open-source communities have intensified their efforts to develop opacs with advanced features. the next-generation catalog is becoming the cur- rent catalog. the library community welcomes open-source integrated library systems (ilss) with open arms, as evi- denced by the increasing number of libraries and library consortia that have adopted or are considering open- source options, such as koha, evergreen, and the open library environment project (ole project). librarians see a golden opportunity to add features to a system that will take years for a proprietary vendor to develop. open-source opacs, especially that of koha, seem to be more innovative than their long-established propri- etary counterparts, as our investigation shows in this paper. threatened by this phenomenon, ils vendors have rushed to improve their opacs, modeling them after the next-generation catalog. for example, ex libris pushed out its new opac, webvoyage . , in august of to give its opac a modern touch. one interesting question remains. in a competition for a modernized opac, which opac is closest to our visions for the next-generation library catalog: open- source or proprietary? the comparative study described in this article was conducted in the hope of yielding some information on this topic. for libraries facing options between open-source and proprietary systems, “a thor- ough process of evaluating an integrated library system (ils) today would not be complete without also weighing the open source ils products against their proprietary counterparts.” ■■ scope and purpose of the study the purpose of the study is to determine which opac of the three ilss—koha, evergreen, or webvoyage—offers more in terms of services and is more comparable to the next-generation library catalog. the three systems include two open-source and one proprietary ilss. koha and evergreen are chosen because they are the two most popular and fully developed open-source ilss in north america. at the time of the study, koha had implementations worldwide; evergreen had library users. we chose webvoyage for comparison because it is the opac of the voyager ils by ex libris, the biggest ils vendor in terms of personnel and marketplace. it also is one of the more popular ilss in north america, with a customer base of , libraries, most of which are academic. as the sample only includes three ilss, the study is very limited in scope, and the findings cannot be extrapolated to all open-source and proprietary cata- logs. but, hopefully, readers will gain some insight into how much progress libraries, vendors, and open-source communities have achieved toward the next-generation catalog. ■■ literature review a review of the library literature found two relevant studies on the comparison of opacs in recent years. the first study was conducted by two librarians in slovenia investigating how much progress libraries had made toward the next-generation catalog. six online catalogs sharon q. yang (yangs@rider.edu) is systems librarian and melissa a. hofmann (mhofmann@rider.edu) is bibliographic control librarian, rider university. information technology and libraries | september were examined and evaluated, including worldcat, the slovene union catalog cobiss, and those of four public libraries in the united states. the study also compared services provided by the library catalogs in the sample with those offered by amazon. the comparison took place primarily in six areas: search, presentation of results, enriched content, user participation, personaliza- tion, and web . technologies applied in opacs. the authors gave a detailed description of the research results supplemented by tables and snapshots of the catalogs in comparison. the findings indicated that “the progress of library catalogues has really been substantial in the last few years.” specifically, the library catalogues have made “the best progress on the content field and the least in user participation and personalization.” when compared to services offered by amazon, the authors concluded that “none of the six chosen catalogues offers the com- plete package of examined options that amazon does.” in other words, library catalogs in the sample still lacked features compared to amazon. the other comparative study was conducted by linda riewe, a library school student, in fulfillment for her master’s degree from san jose university. the research described in her thesis is a questionnaire sur- vey targeted at libraries that compares open-source (specifically, koha and evergreen) and propriety ilss in north america. more than twenty proprietary systems were covered, including horizon, voyager, millennium, polaris, innopac, and unicorn. only a small part of her study was related to opacs. it involved three questions about opacs and asked librarians to evaluate the ease of use of their ils opac’s search engines, their opac search engine’s completeness of features, and their per- ception of how easy it is for patrons to make self-service requests online for renewals and holds. a scale of to was used ( = least satisfied; = very satisfied) regarding the three aspects of opacs. the mean and medium satis- faction ratings for open-source opacs were higher than those of proprietary ones. koha’s opac was ranked . , . , and . , respectively in mean, the highest on the scale in all three categories, while the proprietary opacs were ranked . , . , and . . evergreen fell in the middle, still ahead of proprietary opacs. the findings reinforced the perception that open-source catalogs, especially koha, offer more advanced features than proprietary ones. as riewe’s study focused more on the cost and user satisfac- tion with ilss, it yielded limited information about the connected opacs. no comparative research has measured the progress of open-source versus proprietary catalogs toward the next-generation library catalog. therefore the comparison described in this paper is the first of its kind. as only koha, everygreen, and voyager’s opacs are examined in this paper, the results cannot be extrapolated. studies on a larger scale are needed to shed light on the progress librarians have made toward the next-generation catalog. ■■ method the first step of the study was identifing and defin- ing of a set of measurements by which to compare the three opacs. a review of library literature on the next-generation library catalog revealed different and somewhat conflicting points of views as to what the next- generation catalog should be. as marshall breeding put it, “there isn’t one single answer. we will see a number of approaches, each attacking the problem somewhat dif- ferently.” this study decided to use the most commonly held visions, which are summarized well by breeding and by morgan’s lita executive summary. the ten parameters identified and used in the comparison were taken primarily from breeding’s introduction to the july/ august issue of library technology reports, “next- generation library catalogs.” the ten features reflect some librarians’ visions for a modern catalog. they serve as additions to, rather than replacements of, the feature sets commonly found in legacy catalogs. the following are the definitions of each measurement: ■■ a single point of entry to all library information: “information” refers to all library resources. the next-generation catalog contains not only biblio- graphical information about printed books, video tapes, and journal titles but also leads to the full text of all electronic databases, digital archives, and any other library resources. it is a federated search engine for one-stop searching. it not only allows for one search leading to a federation of results, it also links to full-text electronic books and journal articles and directs users to printed materials. ■■ state-of-the-art web interface: library catalogs should be “intuitive interfaces” and “visually appealing sites” that compare well with other internet search engines. a library’s opac can be intimidating and complex. to attract users, the next-generation catalog looks and feels similar to google, amazon, and other popular websites. this criterion is highly subjective, however, because some users may find google and amazon anything but intuitive or appealing. the underlying assumption is that some internet search engines are popular, and a library catalog should be similar to be popular themselves. ■■ enriched content: breeding writes, “legacy catalogs tend to offer text-only displays, drawing only on the marc record. a next-generation catalog might bring in content from different sources to strengthen the visual appeal and increase the amount of informa- tion presented to the user.” the enriched content the next generation library catalog | yang and hofmann includes images of book covers, cd and movie cases, tables of contents, summaries, reviews, and photos of items that traditionally are not present in legacy catalogs. ■■ faceted navigation: faceted navigation allows users to narrow their search results by facets. the types of facets may include subjects, authors, dates, types of materials, locations, series, and more. many dis- covery tools and federated search engines, such as villanova university’s vufind and innovative interface’s encore, have used this technology in searches. auto-graphics also applied this feature in their opac, agent iluminar. ■■ simple keyword search box: the next-generation catalog looks and feels like popular internet search engines. the best example is google’s simple user interface. that means that a simple keyword search box, instead of a controlled vocabulary or specific-field search box, should be presented to the user on the opening page with a link to an advanced search for user in need of more complex searching options. ■■ relevancy: traditional ranking of search results is based on the frequency and positions of terms in bibliographical records during keyword searches. relevancy has not worked well in opacs. in addi- tion, popularity is another factor that has not been taken into consideration in relevancy ranking. for instance, “when ranking results from the library’s book collection, the number of times that an item has been checked out could be considered an indicator of popularity.” by the same token, the size and font of tags in a tag cloud or the number of comments users attach to an item may also be considered relevant in ranking search results. so far, almost no opacs are capable of incorporating circulation statistics into relevancy ranking. ■■ “did you mean . . . ?”: when a search term is not spelled correctly or nothing is found in the opac in a keyword search, the spell checker will kick in and suggest the correct spelling or recommend a term that may match the user’s intended search term. for exam- ple, a modern catalog may generate a statement such as “did you mean . . . ?” or “maybe you meant . . . .” this may be a very popular and useful service in modern opacs. ■■ recommendations and related materials: the next- generation catalog is envisioned as promoting read- ing and learning by making recommendations of additional related materials to patrons. this feature is an imitation of amazon and websites that promote selling by stating “customers who bought this item also bought . . . .” likewise, after a search in the opac, a statement such as “patrons who borrowed this book also borrowed the following books . . .” may appear. ■■ user contribution—ratings, reviews, comments, and tag- ging: legacy catalogs only allow catalogers to add content. in the next-generation catalog, users can be active contributors to the content of the opac. they can rate, write reviews, tag, and comment on items. user contribution is an important indicator for use and can be used in relevancy ranking. ■■ rss feeds: the next-generation catalog is dynamic because it delivers lists of new acquisitions and search updates to users through rss feeds. modern catalogs are service-oriented; they do more than pro- vide a simple display search results. the second step is to apply these ten visions to the opacs of koha, evergreen, and webvoyage to determine if they are present or absent. the opacs used in this study included three examples from each system. they may have been product demos and live catalogs ran- domly chosen from the user list on the product websites. the latest releases at the time of the study was koha . , evergreen . , webvoyage . . in case of discrepancies between product descriptions and reality, we gave pre- cedence to reality over claims. in other words, even if the product documentation lists and describes a feature, this study does not include it if the feature is not in action either in the demo or live catalogs. despite the fact that a planned future release of one of those investigated opacs may add a feature, this study only recorded what existed at the time of the comparison. the following are the opacs examined in this paper. koha ■■ koho demo for academic libraries: http://academic .demo.kohalibrary.com/ ■■ wagner college: http://wagner.waldo.kohalibrary .com/ ■■ clearwater christian college: http://ccc.kohalibrary .com/ evergreen ■■ evergreen demo: http://demo.gapines.org/opac/ en-us/skin/default/xml/index.xml ■■ georgia pines: http://gapines.org/opac/en-us/ skin/default/xml/index.xml ■■ columbia bible college at http://columbiabc .evergreencatalog.com/opac/en-ca/skin/default/ xml/index.xml webvoyage ■■ rider university libraries: http://voyager.rider.edu ■■ renton college library: http://renton.library.ctc .edu/vwebv/searchbasic information technology and libraries | september ■■ shoreline college library: http://shoreline.library .ctc.edu/vwebv/searchbasic the final step includes data collection and compila- tion. a discussion of findings follows. the study draws conclusions about which opac is more advanced and has more features of the next-generation library catalog. ■■ findings each of the opacs of koha, evergreen, and webvoyage are examined for the presence of the ten features of the next-generation catalog. single point of entry for all library information none of the opacs of the three ilss provides true fed- erated searching. to varying degrees, each is limited in access, showing an absence of contents from elec- tronic databases, digital archives, and other sources that generally are not located in the legacy catalog. of the three, koha is more advanced. while webvoyage and evergreen only display journal-holdings information in their opacs, koha links journal titles from its catalog to proquest’s serials solutions, thus leading users to full- text journals in the electronic databases. the example in figure (koha demo) shows the journal title unix update with an active link to the full-text journal in the availabil- ity field. the link takes patrons to serials solutions, where full text at the journal-title level is listed for each database (see figure ). each link will take you into the full text in each database. state-of-the-art web interface as beauty is in the eye of the beholder, the interface of a catalog can be appealing to one user but prohibitive to another. with this limitation in mind, the out-of-the- box user interface at the demo sites was considered for each opac. all the three catalogs have the google-like simplicity in presentation. all of the user interfaces are highly customizable. it largely depends on the library to make the user interface appealing and welcoming to users. figures – show snapshots from each ilss demo sites and have not been customized. however, there are a few differences in the “state of the art.” for one, koha’s navigation between screens relies solely on the browser’s forward and back buttons, while webvoyage and evergreen have internal naviga- tion buttons that more efficiently take the user between title lists, headings lists, and record displays, and between records in a result set. while all three opacs offer an advanced search page with multiple boxes for entering search terms, only webvoyage makes the relationship between the terms in different boxes clear. by the use of a drop-down box, it makes explicit that the search terms are by default anded and also allows for the selection of or and not. in koha’s and evergreen’s advanced search, however, the terms are anded only, a fact that is not at all obvious to the user. in the demo opacs examined, there is no option to choose or or not between rows, nor is there any indication that the search is anded. the point of providing multiple search boxes is to guide users in constructing a boolean search without their having to worry about operators and syntax. in koha, however, users have to type an or or not statement themselves within the text box, thus defeating the purpose of hav- ing multiple boxes. while evergreen allows for a not construction within a row (“does not contain”), it does not provide an option for or (“contains” and “matches exactly” are the other two options available). see figures figure . link to full-text journals in serials solutions in koha figure . links to serials solutions from koha the next generation library catalog | yang and hofmann – . thus koha’s and evergreen’s advanced search is less than intuitive for users and certainly less functional than webvoyage’s. enriched content to varying degrees, enriched content is present in all three catalogs, with koha providing the most. while all three catalogs have book covers and movie-container art, koha has much more in its catalog. for instance, it displays tags, descriptions, comments, and amazon reviews. webvoyage displays links to google books for book reviews and content summaries but does not have tags, descriptions, and comments in the catalog. see fig- ures – . faceted navigation the koha opac is the only catalog of the three to offer faceted navigation. the “refine your search” feature allows users to narrow search results by availability, places, libraries, authors, topics, and series. clicking on a term within a facet adds that term to the search query and generates a narrower list of results. the user may then choose another facet to further refine the search. while evergreen appears to have faceted navigation upon first glance, it actually does not possess this feature. the following facets appear after a search generates hits: “relevant subjects,” “relevant authors,” and “relevant series.” but choosing a term within a facet does not nar- row down the previous search. instead, it generates an entirely new search with the selected term; it does not add the new term to the previous query. users must manually combine the terms in the simple search box or through the advanced search page. webvoyage also does not offer faceted navigation—it only provides an option to “filter your search” by format, language, and date when a set of results is returned. see figures – . keyword searching koha, evergreen, and webvoyage all present a simple keyword search box with a link to the advanced search (see figures – ). relevancy neither koha, evergreen, nor webvoyage provide any evidence for meeting the criteria of the next-gener- ation catalog’s more inclusive vision of relevancy ranking, such as accounting for an item’s popularity or allowing user tags. koha uses index data’s zebra program for its relevance ranking, which “reads structured records in a variety of input formats . . . and allows access to them through exact boolean search figure . koha: state-of-the-art user interface figure . voyager: state-of-the-art user interface figure . evergreen: state-of-the-art user interface information technology and libraries | september user contributions koha is the only system of the three that allows users to add tags, comments, descriptions, and reviews. in koha’s opac, user-added tags form tag clouds, and the font and size of each keyword or tag indicate that keyword or figure . voyager advanced search figure . koha advanced search figure . evergreen advanced search expressions and relevance-ranked free-text queries. evergreen’s dokuwiki states that the base relevancy score is determined by the cover density of the searched terms. after this base score is determined, items may receive score bumps based on word order, matching on the first word, and exact matches depending on the type of search performed. these statements do not indicate that either koha or evergreen go beyond the traditional relevancy-ranking methods of legacy systems, such as webvoyage. did you mean . . . ? only evergreen has a true “did you mean . . . ?” feature. when no hits are returned, evergreen provides a sug- gested alternate spelling (“maybe you meant . . . ?”) as well as a suggested additional search (“you may also like to try these related searches . . .”). koha has a spell-check feature, but it automatically normalizes the search term and does not give the option of choosing different one. this is not the same as a “did you mean . . . ?” feature as defined above. while the normalizing process may be seamless, it takes the power of choice away from the user and may be problematic if a particular alternative spelling or misspelling is searched purposefully, such as “womyn.” (when “womyn” is searched as a keyword in the koha demo opac, , hits are returned. this catalog does not appear to contain the term as spelled, which is why it is normalized to women. the fact that the term does not appear as is may not be transparent to the searcher.) with normalization, the user may also be unaware that any mistake in spelling has occurred, and the number of hits may differ between the correct spelling and the normalized spelling, potentially affect- ing discovery. the normalization feature also only works with particular combinations of misspellings, where let- ter order affects whether a match is found. otherwise the system returns a “no result found!” message with no suggestions offered. (try “homoexuality” vs. “homo- exsuality.” in koha’s demo opac, the former, with a missing “s,” yields hits, while the latter, with a mis- placed “s,” yields none.) however, koha is a step ahead of webvoyage, which has no built-in spell checker at all. if a search fails, the system returns the message “search resulted in no hits.” see figures – . recommendations/related materials none of the three online catalogs can recommend materi- als for users. the next generation library catalog | yang and hofmann figure . koha enriched content figure . evergreen enriched content figure . voyager enriched content figure . koha faceted navigation figure . evergreen faceted navigation figure . voyager faceted navigation information technology and libraries | september nevertheless, the user contribution in the koha opac is not easy to use. it may take many clicks before a user can figure out how to add or edit text. it requires user login, and the system cannot keep track of the search hits after a login takes place. therefore the user contribution features of koha need improvement. see figure . rss feeds koha provides rss feeds, while evergreen and webvoyage do not. ■■ conclusion table is a summary of the comparisons in this paper. these comparisons show that the koha opac has six out of the ten compared features for the next-generation catalog, plus two halves. its full-fledged features include state-of-the-art web interface, enriched content, faceted navigation, a simple keyword search box, user con- tribution, and rss feeds. the two halves indicate the existence of a feature that is not fully developed. for instance, “did you mean . . . ?” in koha does not work the way the next-generation catalog is envisioned. in addition, koha has the capability of linking journal titles to full text via serials solutions, while the other two opacs only display holdings information. evergreen falls into second place, providing four out of the ten compared features: state-of-the-art interface, enriched content, a keyword search box, and “did you mean . . . ?” webvoyage, the voyager opac from ex libris, comes in third, providing only three out of the ten features for figure . evergreen: did you mean . . . ? figure . koha: did you mean . . . ? figure . voyager: did you mean . . . ? figure . koha user contibutions tag’s frequency of use. all the tags in a tag cloud serve as hyperlinks to library materials. users can write their own reviews to complement the amazon reviews. all user-added reviews, descriptions, and comments have to be approved by a librarian before they are finalized for display in the opac. the next generation library catalog | yang and hofmann the next-generation catalog. based on the evidence, koha’s opac is more advanced and innovative than evergreen’s or voyager’s. among the three catalogs, the open-source opacs compare more favorably to the ideal next-generation catalog then the proprietary opac. however, none of them is capable of federated searching. only koha offers faceted navigation. webvoyage does not even provide a spell checker. the ils opac still has a long way to go toward the next- generation catalog. though this study samples only three catalogs, hopefully the findings will provide a glimpse of the current state of open-source versus proprietary catalogs. ils opacs are not comparable in features and functions to stand-alone opacs, also referred to as “dis- covery tools” or “layers.” some discovery tools, such as ex libris’ primo, also are federated search engines and are modeled after the next-generation catalog. recently they have become increasingly popular because they are bolder and more innovative than ils opacs. two of the best stand-alone open-source opacs are villanova university’s vufind and oregon state university’s libraryfind. both boast eight out of ten features of the next-generation catalog. technically it is easier to develop a new stand-alone opac with all the next-gen- eration catalog features than mending old ils opacs. as more and more libraries are disappointed with their ils opacs, more discovery tools will be implemented. vendors will stop improving ils opacs and concentrate on developing better discovery tools. the fact that ils opacs are falling behind current trends may eventually bear no significance for libraries—at least for the ones that can afford the purchase or implementation of a more sophisticated discovery tool or stand-alone opac. certainly small and public libraries who cannot afford a discovery tool or a programmer for an open-source opac overlay will suffer, unless market conditions change. references . tanja mercun and maja Žumer, “new generation of cata- logues for the new generation of users: a comparison of six library catalogues,” program: electronic library & information systems , no. (july ): – . . eric lease morgan, “a ‘next-generation’ library catalog— executive summary (part # of ),” online posting, july , , lita blog: library information technology association, http:// litablog.org/ / / /a-next-generation-library-catalog -executive-summary-part- -of- / (accessed nov. , ). . marshall breeding, introduction to “next generation library catalogs,” library technology reports , no. (july/aug. ): – . . ibid. . marshall breeding, “library technology guides: key resources in the field of library automation,” http:// www .librarytechnology.org/lwc-search-advanced.pl (accessed jan. , ). . marshall breeding, “investing in the future: automation marketplace ,” library journal (apr. , ), http:// www .libraryjournal.com/article/ca .html (accessed jan. , ). . marshall breeding, “library technology guides: com- pany directory,” http://www.librarytechnology.org/exlibris .pl?sid= &code=vend (accessed jan. , ). . merčun and zumer, “new generation of catalogues.” . ibid. . linda riewe, “integrated library system (ils) survey: open source vs. proprietary-tables” (master’s thesis, san jose university, ): – , http://users.sfo.com/~lmr/ils-survey/ tables-all.pdf (accessed nov. , ). . ibid., – . . breeding, introduction. . ibid.; morgan, “a ‘next-generation’ library catalog.” . breeding, introduction. . ibid. . ibid. . villanova university, “vufind,” http://vufind.org/ (accessed june , ); innovated interfaces, “encore,” http:// encoreforlibraries.com/ (accessed june , ). . auto-graphics, “agent illuminar,” http://www .auto -graphics.com/solutions/agentiluminar/agentiluminar.htm (accessed june , ). . breeding, introduction; morgan, “a ‘next-generation’ table . summary features of the next generation catalog koha evergreen voyager single point of entry for all library information ûü û û state-of-the-art web interface ü ü ü enriched content ü ü ü faceted navigation ü û û keyword search ü ü ü relevancy û û û did you mean…? üû ü û recommended/ related materials û û û user contribution ü û û rss feed ü û û information technology and libraries | september . villanova university, “vufind”; oregon state university, “libraryfind,” http://libraryfind.org/ (accessed june , ). . sharon q.yang and kurt wagner, “open source stand- alone opacs,” (microsoft powerpoint presentation, virtual academic library environment annual conference, piscataway, new jersey, jan. , ). library catalog.” . index data, “zebra,” http://www.indexdata.dk/zebra/ (accessed jan. , ). . evergreen docuwiki, “search relevancy ranking,” http://open-ils.org/dokuwiki/doku.php?id=scratchpad:opac_ demo&s=core (accessed dec. , ). lita cover , cover yalsa cover index to advertisers the next generation library catalog | zhou are your digital documents web friendly? | zhou are your digital documents web friendly?: making scanned documents web accessible the internet has greatly changed how library users search and use library resources. many of them prefer resources available in electronic format over tradi- tional print materials. while many docu- ments are now born digital, many more are only accessible in print and need to be digitized. this paper focuses on how the colorado state university libraries cre- ates and optimizes text-based and digitized pdf documents for easy access, download- ing, and printing. t o digitize print materials, we normally scan originals, save them in archival digital formats, and then make them web- accessible. there are two types of print documents, graphic-based and text-based. if we apply the same tech- niques to digitize these two different types of materials, the documents produced will not be web-friendly. graphic-based materials include archival resources such as his- torical photographs, drawings, manuscripts, maps, slides, and post- ers. we normally scan them in color at a very high resolution to capture and present a reproduction that is as faithful to the original as possible. then we save the scanned images in tiff (tagged image file format) for archival purposes and convert the tiffs to jpeg (joint photographic experts group) or jpeg for web access. however, the same practice is not suitable for modern text-based documents, such as reports, jour- nal articles, meeting minutes, and theses and dissertations. many old text-based documents (e.g., aged newspapers and books), should be yongli zhoututorial files for fast web delivery as access files. for text-based files, access files normally are pdfs that are converted from scanned images. “bcr’s cdp digital imaging best practices version . ” says that the master image should be the highest quality you can afford, it should not be edited or processed for any specific output, and it should be uncom- pressed. this statement applies to archival images, such as photographs, manuscripts, and other image-based materials. if we adopt the same approach for modern text documents, the result may be problematic. pdfs that are created from such master files may have the following drawbacks: ■■ because of their large file size, they require a long download time or cannot be downloaded because of a timeout error. ■■ they may crash a user’s com- puter because they use more memory while viewing. ■■ they sometimes cannot be printed because of insufficient printer memory. ■■ poor print and on-screen view- ing qualities can be caused by background noise and bleed- through of text. background noise can be caused by stains, highlighter marks made by users, and yellowed paper from aged documents. ■■ the ocr process sometimes does not work for high-resolu- tion images. ■■ content creators need to spend more time scanning images at a high resolution and converting them to pdf documents. web-friendly files should be small, accessible by most users, full-text searchable, and have good treated as graphic-based material. these documents often have faded text, unusual fonts, stains, and col- ored background. if they are scanned using the same practice as modern text documents, the document cre- ated can be unreadable and contain incorrect information. this topic is covered in the section “full-text searchable pdfs and troubleshooting ocr errors.” currently, pdf is the file format used for most digitized text docu- ments. while pdfs that are created from high-resolution color images may be of excellent quality, they can have many drawbacks. for exam- ple, a multipage pdf may have a large file size, which increases down- load time and the memory required while viewing. sometimes the down- load takes so long it fails because a time-out error occurs. printers may have insufficient memory to print large documents. in addition, the optical character recognition (ocr) process is not accurate for high- resolution images in either color or grayscale. as we know, users want the ability to easily download, view, print, and search online textual docu- ments. all of the drawbacks created by high-quality scanning defeat one of the most important purposes of digitizing text-based documents: making them accessible to more users. this paper addresses how colorado state university libraries (csul) manages these problems and others as staff create web-friendly digitized textual documents. topics include scanning, long-time archiving, full-text searchable pdfs and troubleshooting ocr problems, and optimizing pdf files for web delivery. preservation master files and access files for digitization projects, we normally refer to images in uncompressed tiff format as master files and compressed yongli zhou is digital repositories librarian, colorado state university libraries, colorado state university, fort collins, colorado information technology and libraries | september information technology and libraries | september factors that determine pdf file size. color images typically generate the largest pdfs and black-and-white images generate the smallest pdfs. interestingly, an image of smaller file size does not necessarily generate a smaller pdf. table shows how file format and color mode affect pdf file size. the source file is a page contain- ing black-and-white text and line art drawings. its physical dimensions are . " by . ". all images were scanned at dpi. csul uses adobe acrobat professional to create pdfs from scanned images. the current ver- sion we use is adobe acrobat professional, but most of its features listed in this paper are available for other acrobat versions. when acrobat converts tiff images to a pdf, it compresses images. therefore a final pdf has a smaller file size than the total size of the original images. acrobat compresses tiff uncom- pressed, lzw, and zip the same amount and produces pdfs of the same file size. because our in-house scanning software does not support tiff g , we did not include tiff g test data here. by comparing simi- lar pages, we concluded that tiff g works the same as tiff uncom- pressed, lzw, and zip. for example, if we scan a text-based page as black- and-white and save it separately in tiff uncompressed, lzw, zip, or g , then convert each page into a pdf, the final pdf will have the same file size without a noticeable quality difference. tiff jpeg generates the smallest pdf, but it is a lossy format, so it is not recommended. both jpeg and jpeg have smaller file sizes but generate larger pdfs than those converted from tiff images. recommendations . use tiff uncompressed or lzw in bits color for pages with color graphs or for historical doc- uments. . use tiff uncompressed or lzw compress an image up to per- cent. some vendors hesitate to use this format because it was proprietary; however, the patent expired on june , . this format has been widely adopted by much software and is safe to use. csul saves all scanned text documents in this format. ■■ tiff zip: this is a lossless compression. like lzw, zip compression is most effective for images that contain large areas of single color. ■■ tiff jpeg: this is a jpeg file stored inside a tiff tag. it is a lossy compression, so csul does not use this file format. other image formats: ■■ jpeg: this format is a lossy com- pression and can only be used for nonarchival purposes. a jpeg image can be converted to pdf or embedded in a pdf. however, a pdf created from jpeg images has a much larger file size com- pared to a pdf created from tiff images. ■■ jpeg : this format’s file extension is .jp . this format offers superior compression per- formance and other advantages. jpeg normally is used for archival photographs, not for text-based documents. in short, scanned images should be saved as tiff files, either with compression or without. we recom- mend saving text-only pages and pages containing text and/or line art as tiff g or tiff lzw. we also recommend saving pages with photo- graphs and illustrations as tiff lzw. we also recommend saving pages with photographs and illustrations as tiff uncompressed or tiff lzw. how image format and color mode affect pdf file size color mode and file format are two on-screen viewing and print quali- ties. in the following sections, we will discuss how to make scanned docu- ments web-friendly. scanning there are three main factors that affect the quality and file size of a digitized document: file format, color mode, and resolution of the source images. these factors should be kept in mind when scanning text documents. file format and compression most digitized documents are scanned and saved as tiff files. however, there are many different formats of tiff. which one is appro- priate for your project? ■■ tiff: uncompressed format. this is a standard format for scanned images. however, an uncom- pressed tiff file has the largest file size and requires more space to store. ■■ tiff g : tiff with g compres- sion is the universal standard for faxs and multipage line-art documents. it is used for black- and-white documents only. ■■ tiff g : tiff with g com- pression has been approved as a lossless archival file format for bitonal images. tiff images saved in this compression have the smallest file size. it is a stan- dard file format used by many commercial scanning vendors. it should only be used for pages with text or line art. many scan- ning programs do not provide this file format by default. ■■ tiff huffmann: a method for compressing bi-level data based on the ccitt group d fac- simile compression schema. ■■ tiff lzw: this format uses a lossless compression that does not discard details from images. it may be used for bitonal, gray- scale, and color images. it may the next generation library catalog | zhou are your digital documents web friendly? | zhou to be scanned at no less than dpi in color. our experiments show that documents scanned at or dpi are sufficient for creating pdfs of good quality. resolutions lower than dpi are not recom- mended because they can degrade image quality and produce more ocr errors. resolutions higher than dpi also are not recommended because they generate large files with little improved on-screen viewing and print quality. we compared pdf files that were converted from images of resolutions at , , and dpi. viewed at percent, the differ- ence in image quality both on screen and in print was negligible. if a page has text with very small font, it can be scanned at a higher resolution to improve ocr accuracy and viewing and print quality. table shows that high-resolu- tion images produce large files and require more time to be converted into pdfs. the time required to combine images is not significantly different compared to scanning time and ocr time, so it was omitted. our example is a modern text docu- ment with text and a black-and-white chart. most of our digitization projects do not require scanning at dpi; dpi is the minimum requirement. we use dpi for most documents and choose a proper color mode for each page. for example, we scan our theses and dissertations in black-and- white at dpi for bitonal pages. we scan pages containing photographs or illustrations in -bit grayscale or -bit color at dpi. other factors that affect pdf file size in addition to the three main fac- tors we have discussed, unnecessary edges, bleed-through of text and graphs, background noise, and blank pages also increase pdf file sizes. figure shows how a clean scan can largely reduce a pdf file size and cover. the updated file has a file size of . mb. the example can be accessed at http://hdl.handle .net/ / . sometimes we scan a page containing text and photo- graphs or illustrations twice, in color or grayscale and in black-and-white. when we create a pdf, we com- bine two images of the same page to reproduce the original appearance and to reduce file size. how to opti- mize pdfs using multiple scans will be discussed in a later section. how image resolution affects pdf file size before we start scanning, we check with our project manager regarding project standards. for some funded projects, documents are required in grayscale bits for pages with black-and-white photographs or grayscale illustrations. . use tiff uncompressed, lzw, or g in black-and-white for pages containing text or line art. to achieve the best result, each page should be scanned accordingly. for example, we had a document with a color cover, pages containing text and line art, and blank pages. we scanned the original document in color at dpi. the pdf created from these images was mb, so large that it exceeded the maximum file size that our repository software allows for uploading. to optimize the document, we deleted all blank pages, converted the pages with text and line art from color to black- and-white, and retained the color table . file format and color mode versus pdf file size file format scan specifications tiff size (kb) pdf size (kb) tiff color bits , tiff lzw color bits , tiff zip color bits , tiff jpeg color bits , jpeg color bits , , jpeg color bits , , tiff grayscale bits , tiff lzw grayscale bits , tiff zip grayscale bits , tiff jpeg grayscale bits , jpeg grayscale bits , , jpeg grayscale bits , , tiff black-and-white tiff lzw black-and-white tiff zip black-and-white note: black-and-white scans cannot be saved in jpeg, jpeg , or tiff jpeg formats. information technology and libraries | september information technology and libraries | september many pdf files cannot be saved as pdf/a files. if an error occurs when saving a pdf to pdf/a, you may use adobe acrobat preflight (advanced > preflight) to identify problems. see figure . errors can be created by non- embedded fonts, embedded images with unsupported file compression, bookmarks, embedded video and audio, etc. by default, the reduce file size procedure in acrobat professional compresses color images using jpeg compression. after running the reduce file size pro- cedure, a pdf may not be saved as a pdf/a because of a “jpeg compression used” error. according to the pdf/a competence center, this problem will be eliminated in the second part of the pdf/a standard— pdf/a- is planned for / . there are many other features in new pdfs; for example, transparency and layers will be allowed in pdf/a- . however, at the time this paper was written pdf/a- had not been announced. portable, which means the file cre- ated on one computer can be viewed with an acrobat viewer on other computers, handheld devices, and on other platforms. a pdf/a document is basically a traditional pdf document that fulfills precisely defined specifications. the pdf/a standard aims to enable the creation of pdf documents whose visual appearance will remain the same over the course of time. these files should be software-independent and unrestricted by the systems used to create, store, and reproduce them. the goal of pdf/a is for long-term archiving. a pdf/a document has the same file extension as a regular pdf file and must be at least compat- ible with acrobat reader . there are many ways to cre- ate a pdf/a document. you can convert existing images and pdf files to pdf/a files, export a doc- ument to pdf/a format, scan to pdf/a, to name a few. there are many software programs you can use to create pdf/a, such as adobe acrobat professional and later ver- sions, compart ag, pdflib, and pdf tools ag. simultaneously improve its viewing and print quality. recommendations . unnecessary edges: crop out. . bleed-through text or graphs: place a piece of white or black card stock on the back of a page. if a page is single sided, use white card stock. if a page is double sided, use black card stock and increase contrast ratio when scanning. often color or grayscale images have bleed- through problems. scanning a page containing text or line art as black-and-white will eliminate bleed-through text and graphs. . background noise: scanning a page containing text or line art as black-and-white can elimi- nate background noise. many aged documents have yellowed papers. if we scan them as color or grayscale, the result will be images with yellow or gray back- ground, which may increase pdf file sizes greatly. we also recom- mend increasing the contrast for better ocr results when scanning documents with background colors. . blank pages: do not include if they are not required. blank pages scanned in grayscale or color can quickly increase file size. pdf and long- term archiving pdf/a pdf vs. pdf/a pdf, short for portable document format, was developed by adobe as a unique format to be viewed through adobe acrobat view- ers. as the name implies, it is table . color mode and image resolution vs. pdf file size color mode resolution (dpi) scanning time (sec.) ocr time (sec.) tiff lzw (kb) pdf size (kb) color n/a* , , color , , color , grayscale , , grayscale , grayscale , b/w b/w b/w *n/a due to an ocr error the next generation library catalog | zhou are your digital documents web friendly? | zhou able. this option keeps the origi- nal image and places an invisible text layer over it. recommended for cases requiring maximum fidelity to the original image. this is the only option used by csul. . searchable image: ensures that text is searchable and selectable. this option keeps the original image, de-skews it as needed, and places an invisible text layer over it. the selection for downs- ample images in this same dia- log box determines whether the image is downsampled and to what extent. the downsam- pling combines several pixels in an image to make a single larger pixel; thus some informa- tion is deleted from the image. however, downsampling does not affect the quality of text or line art. when a proper setting is used, the size of a pdf can be significantly reduced with little or no loss of detail and precision. . clearscan: synthesizes a new type font that closely approxi- mates the original, and preserves the page background using a low-resolution copy. the final pdf is the same as a born-dig- ital pdf. because acrobat can- not guarantee the accuracy of manipulate the pdf document for accessibility. once ocr is properly applied to the scanned files, how- ever, the image becomes searchable text with selectable graphics, and one may apply other accessibility features to the document. acrobat professional provides three ocr options: . searchable image (exact): ensures that text is searchable and select- full-text searchable pdfs and trouble- shooting ocr errors a pdf created from a scanned piece of paper is inherently inaccessible because the content of the docu- ment is an image, not searchable text. assistive technology cannot read or extract the words, users cannot select or edit the text, and one cannot figure . pdfs converted from different images: (a) the original pdf converted from a grayscale image and with unnecessary edges; (b) updated pdf converted from a black- and-white image and with edges cropped out; (c) screen viewed at percent of the pdf in grayscale; and (d) screen viewed at percent of the pdf in black-and-white. dimensions: . ” x . ” color mode: grayscale resolution: dpi tiff lzw: . mb pdf: , kb dimensions: ” x . ” color mode: black-and-white resolution: dpi tiff lzw: kb pdf: kb figure . example of adobe acrobat preflight information technology and libraries | september information technology and libraries | september but at least users can read all text, while the black-and-white scan con- tains unreadable words. troubleshoot ocr error : cannot ocr image based text the search of a digitized pdf is actually performed on its invis- ible text layer. the automated ocr process inevitably produces some incorrectly recognized words. for example, acrobat cannot recognize the colorado state university logo correctly (see figure ). unfortunately, acrobat does not provide a function to edit a pdf file’s invisible text layer. to manu- ally edit or add ocr’d text, adobe acrobat capture . (see figure ) must be purchased. however, our tests show that capture . has many drawbacks. this software is compli- cated and produces it’s own errors. sometimes it consolidates words; other times it breaks them up. in addition, it is time-consuming to add or modify invisible text layers using acrobat capture . . at csul, we manually add searchable text for title and abstract pages only if they cannot be ocr’d by acrobat correctly. the example in troubleshoot ocr error : could not perform recognition (ocr) sometimes acrobat gener- ates an “outside of the allowed specifications” error when process- ing ocr. this error is normally caused by color images scanned at dpi or more. in the example in figure , the page only contains text but was scanned in color at dpi. when we scanned this page as black- and-white at dpi, we did not encounter this problem. we could also use a lower-resolution color scan to avoid this error. our experiments also show that images scanned in black-and-white work best for the ocr process. in this article we mainly discuss running the ocr process on modern textual documents. black-and-white scans do not work well for historical textual documents or aged newspa- pers. these documents may have faded text and background noise. when they are scanned as black- and-white, broken letters may occur, and some text might become unread- able. for this reason they should be scanned in color or grayscale. in fig- ure , images scanned in color might not produce accurate ocr results, ocred text at percent, this option is not acceptable for us. for a tutorial on to how to make a full-text searchable pdf, please see appendix a. troubleshoot ocr error : acrobat crashes occasionally acrobat crashes during the ocr process. the error message does not indicate what causes the crash and where the problem occurs. fortunately, the page number of the error can be found on the top short- cuts menu. in figure , we can see the error occurs on page . we discovered that errors are often caused by figures or diagrams. for a problem like this, the solution is to skip the error-causing page when running the ocr process. our initial research was performed on acrobat professional. our recent study shows that this problem has been significantly improved in acrobat professional. figure . adobe acrobat professional crash window figure . “could not perform recognition (ocr)” error figure . an aged newspaper scanned in color and black-and-white aged newspaper scanned in color aged newspaper scanned in black-and-white the next generation library catalog | zhou are your digital documents web friendly? | zhou a very light yellow background. the undesirable marks and background contribute to its large file size and create ink waste when printed. method : running acrobat’s built-in optimization processes acrobat provides three built-in pro- cesses to reduce file size. by default, acrobat use jpeg compression for color and grayscale images and ccitt group compression for bitonal images. optimize scanned pdf open a scanned pdf and select documents > optimize scanned pdf. a number of settings, such as image quality and background removal, can be specified in the optimize scanned pdf dialog box. our experiments show this process can noticably degrade images and sometimes even increase file size. therefore we do not use this option. reduce file size open a scanned pdf and select documents > reduce file size. the reduce file size command resa- mples and recompresses images, removes embedded base- fonts, and subset-embeds fonts that were left embedded. it also compresses document structure and cleans up elements such as invalid bookmarks. if the file size is already as small as possible, this command has no effect. after process, some files cannot be saved as pdf/a, as we discussed in a previous section. we also noticed that different versions of acrobat can create files of different file sizes even if the same settings were used. pdf optimizer open a scanned pdf and select advanced > pdf optimizer. many settings can be specified in the pdf optimizer dialog box. for example, we can downsample images from sections, we can greatly reduce a pdf’s size by using an appro- priate color mode and resolution. figure shows two different ver- sions of a digitized document. the source document has a color cover and bitonal pages. the origi- nal pdf, shown in figure on the left, was created by another univer- sity department. it was not scanned according to standards and pro- cedures adopted by csul. it was scanned in color at dpi and has a file size of , kb. we exported the original pdf as tiff images, batch-converted color tiff images to black-and-white tiff images, and then created a new pdf using black- and-white tiff images. the updated pdf has a file size of , kb. the image on the right is much cleaner and has better print quality. the file on the left has unwanted marks and figure is a book title page for which we used acrobat capture . to man- ually add searchable text. the entire book may be accessed at http://hdl .handle.net/ / . optimizing pdfs for web delivery a digitized pdf file with color pages may be as large as to mb. most of the time, optimizing processes may reduce files this large without a noticeable difference in quality. in some cases, quality may be improved. we will discuss three optimization methods we use. method : using an appropriate color mode and resolution as we have discussed in previous ~do university original logo text ocred by acrobat figure . incorrectly recognized text sample figure . adobe acrobat capture interface figure . image-based text sample information technology and libraries | september information technology and libraries | september grayscale. a pdf may contain pages that were scanned with different color modes and resolutions. a pdf may also have pages of mixed reso- lutions. one page may contain both bitonal images and color or grayscale images, but they must be of the same resolution. the following strategies were adopted by csul: . combine bitmap, grayscale, and color images. we use gray- scale images for pages that con- tain grayscale graphs, such as black-and-white photos, color images for pages that contain color images, and bitmap images for text-only or text and line art pages. . if a page contains high-definition color or grayscale images, scan that page in a higher resolution and scan other pages at dpi. . if a page contains a very small font and the ocr process does not work well, scan it at a higher resolution and the rest of docu- ment at dpi. . if a page has both text, color, or grayscale graphs, we scan it twice. then we modify images using adobe photoshop and combine two images in acrobat. in figure , the grayscale image has a gray background and a true reproduction of the original photo- graph. the black-and-white scan has a white background and clean text, but details of the photograph are lost. the pdf converted from the grayscale image is kb and has nine ocr errors. the pdf converted from the black-and-white image is kb and has no ocr errors. the pdf converted from a combination of the grayscale and black-and-white images is kb and has no ocr errors. the following are the steps used to create a pdf in figure using acrobat: . scan a page twice—grayscale optimizer can be found at http:// www.acrobatusers.com/tutorials/ understanding-acrobats-optimizer. method : combining different scans many documents have color covers and color or grayscale illustrations, but the majority of pages are text- only. it is not necessary to scan all pages of such documents in color or a higher resolution to a lower reso- lution and choose a different file compression. different collections have different original sources, therefore different settings should be applied. we normally do sev- eral tests for each collection and choose the one that works best for it. we also make our pdfs compat- ible with acrobat to allow users with older versions of software to view our documents. a detailed tutorial of how to use the pdf figure . reduce file size example figure . reduce file size example: combine images the next generation library catalog | zhou are your digital documents web friendly? | zhou help.html?content=wsfd e c b f ea e ab - .html (accessed mar. , ). . ted padova adobe acrobat pdf bible, st ed. (indianapolis: wiley, ). . olaf drümmer, alexandra oettler, and dietrich von seggern, pdf/a in a nutshell—long term archiving with pdf, (berlin: association for digital document standards, ). . pdf/a competence center, “pdf/a: an iso standard—future development of pdf/a,” http://www. pdfa.org/doku.php?id=pdfa:en (accessed july , ). . pdf/a competence center, “pdf/a—a new standard for long- term archiving,” http://www.pdfa.org/ doku.php?id=pdfa:en:pdfa_whitepaper (accessed july , ). . adobe, “creating accessible pdf documents with adobe acrobat . : a guide for publishing pdf documents for use by people with disabilities,” , http://www.adobe.com/enterprise/ a c c e s s i b i l i t y / p d f s / a c ro _ p g _ u e . p d f (accessed mar. , ). . adobe, “recognize text in scanned documents,” , http:// help.adobe.com/en_us/acrobat/ . / s t a n d a rd / w s a d d fa - c fa - c f -b - ab .w.html (accessed mar. , ). . ibid. . ibid. . adobe, “reduce file size by saving,” , http://help.adobe.com/en_us/ acrobat/ . /standard/ws c a -bc c- a - f -b bcd b .w.html (accessed mar. , ). the other pages as grayscale and black-and-white. then we used the procedure described above to com- bine text pages and photographs. the final pdf has clear text and cor- rectly reproduced photographs. the example can be found at http://hdl .handle.net/ / . conclusion our case study, as reported in this article, demonstrates the importance of investing the time and effort to apply the appropriate standards and techniques for scanning and optimiz- ing digitized documents. if proper techniques are used, the final result will be web-friendly resources that are easy to download, view, search, and print. users will be left with a posi- tive impression of the library and feel encouraged to use its materials and services again in the future. references . bcr’s cdp digital imaging best practices working group, “bcr’s cdp digital imaging best practices version . ,” june , http://www.bcr.org/ dps/cdp/best/digital-imaging-bp.pdf (accessed mar. , ). . adobe, “about file formats and compression,” , http://livedocs .adobe.com/en_us/photoshop/ . / and black-and-white. . crop out text on the grayscale scan using photoshop. . delete the illustration on the black-and-white image using photoshop. . create a pdf using the black- and-white image. . run the ocr process and save the file. . insert the color graph. select tools > advanced editing > touchup object tool. right- click on the page and select place image. locate the color graph in the open dialog, then click open and move the color graph to its correct location. . save the file and run the reduce file size or pdf optimizer pro- cedure. . save the file again. this method produces the small- est file size with the best quality, but it is very time-consuming. at csul we used this method for some important documents, such as one of our institutional repository’s show- case items, agricultural frontier to electronic frontier. the book has pages, including a color cover, pages with text and photographs, and text-only pages. we used a color image for the cover page and black-and-white images for the text-only pages. we scanned appendix a. step-by-step creating a full-text searchable pdf in this tutorial, we will show you how to create a full-text searchable pdf using adobe acrobat professional. creating a pdf from a scanner adobe acrobat professional can create a pdf directly from a scanner. acrobat provides five options: black and white document, grayscale document, color document, color image, and custom scan. the custom scan option allows you to scan, run the ocr procedure, add metadata, combine multiple pages into one pdf, and also make it pdf/a compliant. to create a pdf from a scanner, go to file > create pdf > from scanner > custom scan. see figure . at csul, we do not directly create pdfs from scanners because our tests show that it can produce fuzzy text and it is not time efficient. both scanning and running the ocr process can be very time consuming. if an error occurs during these processes, we would have to start over again. we normally scan images on scanning stations by student employees information technology and libraries | september information technology and libraries | september or outsource them to vendors. then library staff will perform quality control and create pdfs on seperate machines. in this way, we can work on multiple documents at the same time and ensure that we provide high-quality pdfs. creating a pdf from scanned images . from the task bar select combine > merge files into a single pdf > from multiple files. see figure . . in the combine files dialog, make sure the single pdf radio button is selected. from the add files dropdown menu select add files. see figure . . in the add files dialog, locate images and select multiple images by holding shift key, and then click add files button. . by default, acrobat sorts files by file names. use move up and move down buttons to change image orders and use the remove button to delete images. choose a target file size. the smallest icon will produce a file with a smaller file size but a lower image quality pdf, and the largest icon will produce a high image quality pdf but with a very large file size. we normally use the default file size setting, which is the middle icon. . save the file. at this point, the pdf is not full-text searchable. making a full-text searchable pdf a pdf document created from a scanned piece of paper is inherently inaccessible because the content of the document is an image, not searchable text. assistive technology cannot read or extract the words, users cannot select or edit the text, and one cannot manipulate the pdf document for accessibility. once optical character recognition (ocr) is properly applied to the scanned files, however, the image becomes searchable text with selectable graphics, and one may apply other acces- sibility features to the document. adobe acrobat professional provides three ocr options, searchable image (exact), searchable image, and clean scan. because searchable image (exact) is the only option that keeps the original look, we only use this option. to run an ocr procedure using acrobat professional: . open a digitized pdf. . select document > ocr text recognition > recognize text using ocr. . in the recognize text dialog, specify pages to be ocred. . in the recognize text dialog, click the edit button in the settings sec- tion to choose ocr language and pdf output style. we recommend the searchable image (exact) option. click ok. the setting will be remembered by the program and will be used until a new setting is chosen. sometimes a pdf’s file size increases greatly after an ocr process. if this happens, use the pdf optimizer to reduce its file size. figure . merge files into a single pdf figure . combine files dialog figure . acrobat professional’s create pdf from scanner dialog editorial | truitt a recent library journal (lj) story referred to “the pal- pable hunger public librarians have for change . . . and, perhaps, a silver bullet to ensure their future” in the context of a presentation at the public library association’s annual conference by staff members of the rangeview (colo.) library district. now, lest there be any doubt on this point, allow me to state clearly from the outset that none of the following ramblings are in any way intended as a specific critique of the measures under- taken by rangeview. far be it from me to second-guess the rangeview staff’s judgment as to how best to serve the community there. rather, what got my attention was lj’s reference to a “palpable hunger”for magic ammunition, from whose presumed existence we in libraries seem to draw com- fort. in the last quarter century, it seems as though we’ve heard about and tried enough silver bullets to keep our collective six-shooters endlessly blazing away. here are just a few examples that i can recall off the top of my head, and in no particular order: ■■ library cafes and coffee shops. ■■ libraries arranged along the lines of chain book- stores. ■■ general-use computers in libraries (including infor- mation/knowledge commons and what-have-you) ■■ computer gaming in libraries. ■■ lending laptops, digital cameras, mp players and ipods, e-book readers, and now ipads. ■■ mobile technology (e.g., sites and services aimed at and optimized for iphones, blackberries, etc.) ■■ e-books and e-serials. ■■ chat and instant-message reference. ■■ libraries and social networking (e.g., facebook, twitter, second life, etc.). ■■ “breaking down silos,” and “freeing”/exposing our bibliographic data to the web, and reuse by others outside of the library milieu. ■■ ditching our old and “outmoded” systems, whether the object of our scorn is aacr , lcsh, lcc, dewey, marc, the ils, etc. ■■ library websites generally. remember how every- one—including us—simply had to have a website in the s? and ever since then, it’s been an endless treadmill race to find the perfect, user-centric library web presence? if sisyphus were to be incarnated today, i have little doubt that he would appear as a library web manager and his boulder would be a library website. ■■ oh, and as long as we’re at it, “user-centricity” gen- erally. the implication, of course, is that before the term came into vogue, libraries and librarians were not focused on users. ■■ “next-gen” catalogs. i’m sure i’m forgetting a whole lot more. anyway, you get the picture. each of these has, at one time or another, been posi- tioned by some advocate as the necessary change—the “silver bullet”—that would save libraries from “irrel- evance” (or worse!), if we would but adopt it now, or better yet, yesterday. well, to judge from the generally dismal state of libraries as depicted by some opinion- makers in our profession—or perhaps simply from our collective lack of self-esteem—we either have been misled about the potency of our ammunition, or else we’ve been very poor markspersons. notwithstanding the fact that we seem to have been indiscriminately blasting away with shotguns rather than six-shooters, our shooting has neither reversed the trends of shrinking budgets and declining morale nor staunched the ceaseless dire warn- ings of some about “irrelevance” resulting from ebbing library use. to stretch the analogy a bit further still, one might even argue that all this shooting has done damage of its own, peppering our most valuable services with countless pellet-sized holes. at the same time, we have in recent years shown ourselves to be remarkably susceptible to the marketing- focused hyperbole of those in and out of librarianship about technological change. each new technology is labeled a “game-changer”; change in general is either— to use the now slightly-dated, oh-so-nineties term—a “paradigm shift” or, more recently, “transformational.” when did we surrender our skepticism and awareness of a longer view? what’s wrong with this picture? i’d like to suggest another way of viewing this. a couple of years ago, alan weisman published the world without us, a book that should be required reading for all who are interested in sustainability, our own hubris, and humankind’s place in the world. the book begins with our total, overnight disappearance, and asks ( ) what would the earth be like without us? and ( ) what evidence of our works would remain, and for how long? the bottom line answers for weisman are ( ) in the long run, probably much better off, and ( ) not much and not for very long, really. so, applying weisman’s first question to our own, much more modest domain, what might the world be like if tomorrow librarians all disappeared or went on to work doing something else—became consultants, perhaps?— and our physical and virtual collections were padlocked? would everything be okay, because as some believe, marc truitteditorial: no more silver bullets, please marc truitt (marc.truitt@ualberta.ca) is associate university librarian, bibliographic and information technology services, university of alberta libraries, edmonton, alberta, canada, and editor of ital. information technology and libraries | june think we need to be prepared to turn off the lights, lock the doors, and go elsewhere, because i hope that what we’re doing is about more than just our own job security. and if the far-fetched should actually happen, and we all disappear? i predict that at some future point, some- one will reinvent libraries and librarians, just as others have reinvented cataloguing in the guise of metadata. notes and references . norman oder, “pla conference: the anythink revolution is ripe,” library journal, mar. , , http://www .libraryjournal.com/article/ca .html (accessed mar. , ). there, i said it! a fairly innocuous disclaimer added to one of my columns last year seemed to garner more attention (http:// freerangelibrarian.com/ / / /marc-truitts-surprising -ital-editorial/) than did the content of the column itself. will the present disclaimer be the subject of similar speculation? . one of my favorite antidotes to such bloated, short-term language is embodied in michael gorman’s “human values in a technological age,” ital , no. (mar. ): – , http:// www.ala.org/ala/mgrps/divs/lita/ital/ gorman.cfm (accessed apr , )—highly recommended. the following is but one of many calming and eminently sensible observations gorman makes: the key to understanding the past is the knowledge that people then did not live in the past—they lived in the present, just a different present from ours. the present we are living in will be the past sooner than we wish. what we perceive as its uniqueness will come to be seen as just a part of the past as viewed from the point of a future present that will, in turn, see itself as unique. people in history did not wear quaintly old- fashioned clothes—they wore modern clothes. they did not see themselves as comparing unfavorably with the people of the future, they compared themselves and their lives favorably with the people of their past. in the context of our area of interest, it is particularly interesting to note that people in history did not see themselves as technologically primitive. on the contrary, they saw themselves as they were—at the leading edge of technology in a time of unprecedented change. it’s all out there on the web anyway, and google will make it findable? absent a few starry-eyed bibliophiles and newly out-of-work librarians—those who didn’t make the grade as consultants—would anyone mourn our disappearance? would anyone notice? if a tree falls in the woods . . . in short, would it matter? and if so, why and how much? the answer to the preceding two questions, i think, can help to point the way to an approach for understand- ing and evaluating services and change in libraries that is both more realistic and less draining than our obsessive quest for the “silver bullet.” what exactly is our “value- add”? what do we provide that is unique and valuable? we can’t hope to compete with barnes and noble, starbucks, or the googleplex; seeking to do so simply diverts resources and energy from providing services and resources that are uniquely ours. instead, new and changed services and approaches should be evaluated in terms of our value-add: if they contribute positively and are within our abilities to do them, great. if they do not contribute positively, then try- ing to do them is wasteful, a distraction, and ultimately disillusioning to those who place their hopes in such panaceas. some of the “bullets” i listed above may well qualify as contributing to our value-add, and that’s fine. my point isn’t to judge whether they are “bad” or “good.” my argument is about process and how we decide what we should do and not do. understanding what we contribute that is uniquely ours should be the reference standard by which proposed changes are evaluated, not some pie-in- the-sky expectation that pursuit of this or that vogue will magically solve our funding woes, contribute to higher (real or virtual) gate counts, make us more “relevant” to a particular user group, or even raise our flagging self- esteem. in other words, our value-add must stand on its own, regardless of whether it actually solves temporal problems. it is the “why” in “why are we here?” if, at the end of the day, we cannot articulate that which makes us uniquely valuable—or if society as a whole finds that contribution not worth the cost—then i editorial board thoughts: ital . | boze litablog.org/) i see that there are occasional posts, but there are rarely comments and little in the way of real dis- cussion. it seems to be oriented toward announcements, so perhaps it’s not a good comparison with italica. some ala groups are using wordpress for their blogs, a few with user comments, but mostly without much apparent traffic (for example, the ll&m online blog, http://www .lama.ala.orgllandm). in general, blogs don’t seem to be a satisfactory platform for discussion. wikis aren’t par- ticularly useful in this regard, either, so i think that rules out the lita wiki (http://wikis.ala.org/lita/index.php/ main_page). i’ve looked at ala connect (http://connect. ala.org/), which has a variety of web . features, so it might be a good home for italica. we could also use a mailing list, either one that already exists, such as lita-l, or a new one. the one advantage e-mail has is that it is delivered to the reader, so one doesn’t have to remember to visit a website. we already have rss feeds for the italica blog, so maybe that works well enough as a notification for those who subscribe to them. i’ve also wondered whether a discussion forum (aka message board) would be useful. i frequent a few soft- ware-related forums, and i find them conducive to discussion. they have a degree of flexibility lacking in other platforms. it’s easy for any participant to start up a new topic rather than limiting discussion only to topics posted by the owner, as is usually the case with blogs. frankly i’d like to encourage discussion on topics beyond only the articles published in ital. for example, we used to have columns devoted to book and software reviews. even though they were discontinued, those could still be interesting topics for discussion between ital readers. in writing this, my hope is to get feedback from you, the reader, about what ital and italica could be doing for you. how can we use ala connect in ways that would be useful? could we use other platforms to do things beyond simply discussing articles that appear in the print edition? what social web technologies do you use, and how could we apply them to ital? after you read this, i hope you’ll join us at italica for a discussion. let us know what you think. editor’s note: andy serves on the ital editorial board and as the ital website manager. he earns our gratitude every quarter with his timely and professional work to post the new issue online. t he title of this recurring column is “editorial board thoughts,” so as i sit here in the middle of february, what am i thinking about? as i trudge to work each day through the snow and ice, i think about what a nuisance it is to have a broken foot (i broke the fifth metatarsal of my left foot at the midwinter meeting in boston—not recommended) but most recently i’ve been thinking about ital. the march issue is due to be mailed in a couple of weeks, and i got the digital files a week or so ago. in a few days i’ll have to start separating the pdf into individ- ual articles, and then i’ll start up my web editor to turn the rtf files for each article into nicely formatted html. all of this gets fed into ala’s content management sys- tem, where you can view it online by pointing your web browser to http://www.lita.org/ala/mgrps/divs/lita/ ital/italinformation.cfm. in case you didn’t realize it, the full text of each issue of ital is there, going back to early . selected full-text articles are available from earlier issues going back to . the site is in need of a face lift, but we expect to work on that in the near future. starting with the september issue of ital we launched italica, the ital blog at http://ital-ica .blogspot.com/, as a pilot. italica was conceived as a forum for readers, authors, and editors of ital to discuss each issue. for a year and a half we’ve been open for reader feedback, and our authors have been posting to the blog and responding to reader comments. what’s your opinion of italica? is it useful? what could we be doing to enhance its usefulness? in reality we haven’t had a great deal of communica- tion via the blog. we are looking at moving italica from blogger to a platform more integrated with existing ala or lita services. is a blog format the best way to encour- age discussion? when i look at the lita blog (http:// andy boze (boze. @nd.edu) is head, desktop computing and network services, university of notre dame hesburgh libraries, notre dame, indiana. andy bozeeditorial board thoughts: ital . information technology and libraries | june tinuing education opportunities for library informa- tion technologists and all library staff who have an interest in technology. . innovation: to serve the library community, lita expert members will identify and demonstrate the value of new and existing technologies within ala and beyond. . advocacy and policy: lita will advocate for and participate in the adoption of legislation, policies, technologies, and standards that promote equitable access to information and technology. . the organization: lita will have a solid structure to support its members in accomplishing its mission, vision, and strategic plan. . collaboration and outreach: lita will reach out and col- laborate with other library organizations to increase the awareness of the importance of technology in libraries, improve services to existing members, and reach out to new members. the lita executive committee is currently finalizing the strategies lita will pursue to achieve success in each of the goal areas. it is my hope that the strategies for each goal are approved by the lita board of directors before the ala annual conference in washington, d.c. that way the finalized version of the lita strategic plan can be introduced to the committee and interest group chairs and the membership as a whole at that conference. this will allow us to start the next fiscal year with a clear road for the future. while i am excited about what is next, i have also been dreading the end of my presidency. i have truly enjoyed my experience as lita president, and in some way wish it was not about to end. i have learned so much and have met so many wonderful people. thank you for giving me this opportunity to serve you and for your support. i have truly appreciated it. a s i write this last column, the song “my way” by frank sinatra keeps going through my head. while this is definitely not my final curtain, it is the final curtain of my presidency. like sinatra i have a few regrets, “but then again, too few to mention.” there was so much more i wanted to accomplish this year; however, as usual, my plans were more ambitious than the time i had available. being lita’s president was a big part of my life, but it was not the only part. those other parts—like family, friends, work, and school—demanded my atten- tion as well. i have thought about what to say in this final column. do i list my accomplishments of the last year? nah, you can read all about that in the lita annual report, which i will post in june. tackle some controversial topic? while i can think of a few, i have not yet thought of any solutions, and i do not want to rant against something without pro- posing some type of solution or plan of attack. i thought instead i would talk about where i have devoted a large part of my lita time over the last year. as i look back at the last year, i am also thinking ahead to the future of lita. we are currently writing lita’s strategic plan. we have a lot to great ideas to work with. lita members are always willing to share their thoughts both formally and informally. i have been charged with the task of taking all of those great ideas, gathered at conferences, board meetings, hallway con- versations, surveys, e-mail, etc., to create a roadmap for the future. after reviewing all of the ideas gathered over the last three years, i was able to narrow that list down to six major goal areas. with the assistance of the lita board of directors and the lita executive committee, we whittled the list down to five major goal areas of the lita strategic plan: . training and continuing education: lita will be nationally recognized as the leading source for con- michelle frisque (mfrisque@northwestern.edu) is lita president – and head, information systems, north- western university, chicago. michelle frisque president’s message: the end and new beginnings web services and widgets for library information systems | han on the clouds: a new way of computing | han shape cloud computing. for exam- ple, sun’s well-known slogan “the network is the computer” was estab- lished in late s. salesforce.com has been providing on-demand software as a service (saas) for cus- tomers since . ibm and microsoft started to deliver web services in the early s. microsoft’s azure service provides an operating sys- tem and a set of developer tools and services. google’s popular google docs software provides web-based word-processing, spreadsheet, and presentation applications. google app engine allows system devel- opers to run their python/java applications on google’s infrastruc- ture. sun provides $ per cpu hour. amazon is well-known for provid- ing web services such as ec and s . yahoo! announced that it would use the apache hadoop frame- work to allow users to work with thousands of nodes and petabytes ( million gigabytes) of data. these examples demonstrate that cloud computing providers are offer- ing services on every level, from hardware (e.g., amazon and sun), to operating systems (e.g., google and microsoft), to software and ser- vice (e.g., google, microsoft, and yahoo!). cloud-computing provid- ers target a variety of end users, from software developers to the general public. for additional infor- mation regarding cloud computing models, the university of california (uc) berkeley’s report provides a good comparison of these models by amazon, microsoft, and google. as cloud computing providers lower prices and it advancements remove technology barriers—such as virtualization and network band- width—cloud computing has moved into the mainstream. gartner stated, “organizations are switching from factors related to cloud computing: infinite computing resources avail- able on demand, removing the need to plan ahead; the removal of an up-front costly investment, allowing companies to start small and increase resources when needed; and a system that is pay-for-use on a short-term basis and releases customers when needed (e.g., cpu by hour, storage by day). national institute of standards and technology (nist) currently defines cloud computing as “a model for enabling convenient, on-demand network access to a shared pool of configurable computing resources (e.g. network, servers, storage, appli- cations, and services) that can be rapidly provisioned and released with minimal management effort or service provider interaction.” as there are several definitions for “utility computing” and “cloud computing,” the author does not intend to suggest a better definition, but rather to list the characteristics of cloud computing. the term “cloud computing” means that ■■ customers do not own network resources, such as hardware, software, systems, or services; ■■ network resources are provided through remote data centers on a subscription basis; and ■■ network resources are delivered as services over the web. this article discusses using cloud computing on an it-infrastructure level, including building virtual server nodes and running a library’s essen- tial computer systems in remote data centers by paying a fee instead of run- ning them on-site. the article reviews current cloud computing services, presents the author’s experience, and discusses advantages and disadvan- tages of using the new approach. all kinds of clouds major it companies have spent bil- lions of dollars since the s to on the clouds: a new way of computing this article introduces cloud computing and discusses the author’s experience “on the clouds.” the author reviews cloud computing services and providers, then presents his experience of running mul- tiple systems (e.g., integrated library sys- tems, content management systems, and repository software). he evaluates costs, discusses advantages, and addresses some issues about cloud computing. cloud com- puting fundamentally changes the ways institutions and companies manage their computing needs. libraries can take advan- tage of cloud computing to start an it project with low cost, to manage computing resources cost-effectively, and to explore new computing possibilities. s cholarly communication and new ways of teaching provide an opportunity for academic institutions to collaborate on pro- viding access to scholarly materials and research data. there is a grow- ing need to handle large amounts of data using computer algorithms that presents challenges to libraries with limited experience in handling nontextual materials. because of the current economic crisis, aca- demic institutions need to find ways to acquire and manage computing resources in a cost-effective manner. one of the hottest topics in it is cloud computing. cloud computing is not new to many of us because we have been using some of its services, such as google docs, for years. in his latest book, the big switch: rewiring the world, from edison to google, carr argues that computing will go the way of electricity: purchase when needed, which he calls “utility computing.” his examples include amazon’s ec (elastic computing cloud), and s (simple storage) services. amazon’s chief technol- ogy officer proposed the following yan hantutorial yan han (hany@u.library.arizona.edu) is associate librarian, university of arizona libraries, tucson. information technology and libraries | june information technology and libraries | june company-owner hardware and software to per-use service-based models.” for example, the u.s. gov- ernment website (http://www.usa .gov/) will soon begin using cloud computing. the new york times used amazon’s ec and s services as well as a hadoop application to pro- vide open access to public domain articles from to . the times loaded tb of raw tiff images and their derivative million pdfs into amazon’s s in twenty-four hours at very reasonable cost. this project is very similar to digital library proj- ects run by academic libraries. oclc announced its movement of library management services to the web. it is clear that oclc is going to deliver a web-based integrated library sys- tem (ils) to provide a new way of running an ils. duraspace, a joint organization by fedora commons and dspace foundation, announced that they would be taking advan- tage of cloud storage and cloud computing. on the clouds computing needs in academic librar- ies can be placed into two categories: user computing needs and library goals. user computing needs academic libraries usually run hun- dreds of pcs for students and staff to fulfill their individual needs (e.g., microsoft office, browsers, and image-, audio-, and video-processing applications). library goals a variety of library systems are used to achieve libraries’ goals to sup- port research, learning, and teaching. these systems include the following: ■■ library website: the website may be built on simple html web- pages or a content management system such as drupal, joomla, or any home-grown php, perl, asp, or jsp system. ■■ ils: this system provides tra- ditional core library work such as cataloging, acquisition, reporting, accounting, and user management. typical systems include innovative interfaces, sirsidynix, voyager, and open- source software such as koha. ■■ repository system: this sys- tem provides submission and access to the institution’s digi- tal collections and scholarship. typical systems include dspace, fedora, eprints, contentdm, and greenstone. ■■ other systems: for example, fed- erated search systems, learning object management systems, interlibrary loan (ill) systems, and reference tracking systems. ■■ public and private storage: staff file-sharing, digitization, and backup. due to differences in end users and functionality, most systems do not use computing resources equally. for example, the ils is input and output intensive and database query intensive, while repository systems require storage ranging from a few gigabytes to dozens of terabytes and substantial network bandwidth. cloud computing brings a funda- mental shift in computing. it changes the way organizations acquire, configure, manage, and maintain computing resources to achieve their business goals. the availability of cloud computing providers allows organizations to focus on their busi- ness and leave general computing maintenance to the major it compa- nies. in the fall of , the author started to research cloud computing providers and how he could imple- ment cloud computing for some library systems to save staff and equipment costs. in january , the author started his plan to build library systems “on the clouds.” the university of arizona libraries (ual) has been a key player in the process of rebuilding higher education in afghanistan since . ual librarian atifa rawan and the author have received multiple grant contracts to build technical infra- structures for afghanistan’s academic libraries. the technical infrastructure includes the following: ■■ afghanistan ils: a bilingual ils based on the open-source system koha. ■■ afghanistan digital libraries website (http://www.afghan digitallibraries.org/): originally built on simple html pages, later rebuilt in using the con- tent management system joomla. ■■ a digitization management sys- tem. the author has also developed a japanese ill system (http://gif project.libraryfinder.org) for the north american coordinating council on japanese library resources. these systems had been running on ual’s internal technical infrastructure. these systems run in a complex computing environment, require different modules, and do not use computing resources equally. for example, the afghan ils runs on linux, apache, mysql, and perl. its opac and staff interface run on two different ports. the afghanistan digital libraries website requires linux, apache, mysql, and php. the japanese ill system was written in java and runs on tomcat. there are several reasons why the author moved these systems to the new cloud computing infrastructure: ■■ these systems need to be accessed in a system mode by people who are not ual employees. ■■ system rebooting time can be substantial in this infrastructure because of server setup and it policy. ■■ the current on-site server has web services and widgets for library information systems | han on the clouds: a new way of computing | han reached its life expectancy and requires a replacement. by analyzing the complex needs of different systems and considering how to use resources more effec- tively, the author decided to run all the systems through one cloud computing provider. by comparing the features and the costs, linode (http://www.linode.com/) was chosen because it provides full ssh and root access using virtualization, four data centers in geographically diverse areas, high availability and clustering support, and an option for month-to-month contracts. in addition, other customers have pro- vided positive reviews. in january , the author purchased one node located in fremont, california, for $ . per month. an imple- mentation plan (see appendix) was drafted to complete the project in phases. the author owns a virtual server and has access to everything that a physical server provides. in addition, the provider and the user community provided timely help and technical support. the migration of systems was straightforward: a linux kernel (debian . ) was installed within an hour, domain registration was com- plete and the domains went active in twenty-four hours, the afghanistan digital libraries’ website (based on joomla) migration was complete within a week, and all supporting tools and libraries (e.g., mysql, tomcat, and java sdk) were installed and configured within a few days. a month later, the afghanistan ils (based on koha) migration was com- pleted. the ill system was also migrated without problem. tests have been performed in all these systems to verify their usabil- ity. in summary, the migration of systems was very successful and did not encoun- ter any barriers. it addresses the issues facing us: after the migration, ssh log-ins for users who are not univer- sity employees were set up quickly; systems maintenance is managed by the author’s team, and rebooting now only takes about one minute; and there is no need to buy a new server and put it in a temperature and security controlled environment. the hardware is maintained by the provider. the administrative gui for the linux nodes is shown in figure . since migration, no downtime because of hardware or other failures caused by the provider has been observed. after migrating all the sys- tems successfully and running them in a reliable mode for a few months, the second phase was implemented (see appendix). another linux node (located in atalanta, georgia) was purchased for backup and moni- toring (see figure ). nagios, an open-source monitoring system, was tested and configured to identify and report problems for the above library systems. nagios provides the follow- ing functions: ( ) monitoring critical computing components, such as the network, systems, services, and serv- ers; ( ) timely alerts delivered via e-mail or cell phone; and ( ) report and record logs of outages, events, and alerts. a backup script is also run as a prescheduled job to back up the systems on a regular basis. figure . linux node administration web interface figure . two linux nodes located in two remote data centers node : . .xxx.xxx (fremont, ca) node : . .xxx.xxx (atlanta, ga) nagios backup afghan digital libraries website afghan ils interlibrary loan system dspace information technology and libraries | june information technology and libraries | june findings and discussions since january , all the systems have been migrated and have been running without any issues caused by the provider. the author is very satisfied with the outcomes and cost. the annual cost of running two nodes is $ per year, compared to at least $ , dollars if the hardware had been run in the library. from the author ’s experience, cloud computing provides the following advantages over the tradi- tional way of computing in academic institutions: ■■ cost-effectiveness: from the above example and literature review, it is obvious that using cloud computing to run applications, systems, and it infrastruc- ture saves staff and financial resources. uc berkeley’s report and zawodny’s blog provide a detailed analysis of costs for cpu hours and disk storage. ■■ flexibility: cloud computing allows organizations to start a project quickly without worrying about up-front costs. computing resources such as disk storage, cpu, and ram can be added when needed. in this case, the author started on a small scale by purchasing one node and added additional resources later. ■■ data safety: organizations are able to purchase storage in data centers located thousands of miles away, increasing data safety in case of natural disasters or other factors. this strategy is very difficult to achieve in a tra- ditional off-site backup. ■■ high availability: cloud comput- ing providers such as microsoft, google, and amazon have bet- ter resources to provide more up-time than almost any other organizations and companies do. ■■ the ability to handle large amounts of data: cloud computing has a pay-for-use business model that allows academic institutions to analyze terabytes of data using distributed computing over hundreds of computers for a short-time cost. on-demand data storage, high availability and data safety are criti- cal features for academic libraries. however, readers should be aware of some technical and business issues: ■■ availability of a service: in sev- eral widely reported cases, amazon’s s and google gmail were inaccessible for a duration of several hours in . the author believes that the com- mercial providers have better technical and financial resources to keep more up-time than most academic institutions. for those wanting no single point of fail- ure (e.g., a provider goes out of business), the author suggests storing duplicate data with a dif- ferent provider or locally. ■■ data confidentiality: most aca- demic libraries have open-access data. this issue can be solved by encrypting data before moving to the clouds. in addition, licens- ing terms can be negotiated with providers regarding data safety and confidentiality. ■■ data transfer bottlenecks: accessing the digital collections requires considerable network bandwidth, and digital collections are usually optimized for customer access. moving huge amounts of data (e.g., preservation digital images, audios, videos, and data sets) to data centers can be scheduled during off hours (e.g., – a.m.), or data can be shipped on hard disks to the data centers. ■■ legal jurisdiction: legal jurisdic- tion creates complex issues for both providers and end users. for example, canadian privacy laws regulate data privacy in public and private sectors. in , the office of the privacy commissioner of canada released a finding that “outsourcing of canada .com email services to u.s.-based firm raises questions for subscrib- ers,” and expressed concerns about public sector privacy pro- tection. this brings concerns to both providers and end users, and it was suggested that privacy issues will be very challenging. summary the author introduces cloud comput- ing services and providers, presents his experience of running multiple sys- tems such as ils, content management systems, repository software, and the other system “on the clouds” since january . using cloud comput- ing brings significant cost savings and flexibility. however, readers should be aware of technical and business issues. the author is very satisfied with his experience of moving library systems to cloud computing. his experience demonstrates a new way of managing critical computing resources in an aca- demic library setting. the next steps include using cloud computing to meet digital collections’ storage needs. cloud computing brings fun- damental changes to organizations managing their computing needs. as major organizations in library fields, such as oclc, started to take advan- tage of cloud computing, the author believes that cloud computing will play an important role in library it. acknowledgments the author thanks usaid and washington state university for pro- viding financial support. the author thanks matthew cleveland’s excel- lent work “on the clouds.” references . nicholars carr, the big switch: rewiring the world, from edison to google web services and widgets for library information systems | han on the clouds: a new way of computing | han (london: norton, ). . werner vogels, “a head in the clouds—the power of infrastructure as a service” (paper presented at the cloud computing and in applications confer- ence (cca ’ ), chicago, oct. – , ). . peter mell and tim grance, “draft nist working definition of cloud com- puting,” national institute of standards and technology (may , ), http:// csrc.nist.gov/groups/sns/cloud-com- puting/index.html (accessed july , ). . michael armbust et al., “above the clouds: a berkeley view of cloud com- puting,” technical report, university of california, berkeley, eecs department, feb. , , http://www.eecs.berkeley .edu/pubs/techrpts/ /eecs- - .html (accessed july , ). . eric hand, “head in the clouds: ‘cloud computing’ is being pitched as a new nirvana for scientists drowning in data. but can it deliver?” nature , no. ( ): ; geoffery fowler and ben worthen, “the internet indus- try is on a cloud—whatever that may mean,” wall street journal, mar. , , http://online.wsj.com/article/ sb .html (accessed july , ); stephen baker, “google and the wisdom of the clouds,” business week (dec. , ), http://www.msnbc .msn.com/id/ / (accessed july , ). . gartner, “gartner says worldwide it spending on pace to supass $ . tril- lion in ,” press release, aug. , , http://www.gartner.com/it/page .jsp?id= (accessed july , ). . wyatt kash, “usa.gov, gobierno usa.gov move into the internet cloud,” government computer news, feb. , , http://gcn.com/articles/ / / / gsa-sites-to-move-to-the-cloud.aspx?s =gcndaily_ (accessed july , ). . derek gottfrid, “self-service, prorated super computing fun!” online posting, new york times open, nov. , , http://open.blogs .nytimes.com/ / / /self-service -prorated-super-computing-fun/?scp = &sq=self% service% prorated&st =cse (accessed july , ). . oclc online computing library center, “oclc announces strategy to move library management services to web scale,” press release, apr. , , http://www.oclc.org/us/en/news/ releases/ .htm (accessed july , ). . duraspace, “fedora commons and dspace foundation join together to create duraspace organization,” press release, may , , http:// duraspace.org/documents/pressrelease .pdf (accessed july , ). . yan han and atifa rawan, “afghanistan digital library initiative: revitalizing an integrated library sys- tem,” information technology & libraries , no. ( ): – . . fowler and worthen, “the internet industry is on a cloud.” . jeremy zawodney, “replacing my home backup server with amazon’s s ,” online posting, jeremy zawod- ny’s blog, oct. , , http://jeremy .zawodny.com/blog/archives/ .html (accessed june , ). . yan han, “an integrated high availability computing platform,” the electronic library , no. ( ): – . . office of the privacy commissioner of canada, “tabling of privacy com- missioner of canada’s – annual report on the privacy act: commissioner expresses concerns about public sector privacy protection,” press release, june , , http://www.priv.gc.ca/media/ nr-c/ /nr-c_ _e.cfm (accessed july , ); office of the privacy com- missioner of canada, “findings under the personal information protection and elec- tronic documents act (pipeda),” (sept. , ), http://www.priv.gc.ca/cf -dc/ / _ _e.cfm (accessed july , ). . stephen baker, “google and the wisdom of the clouds,” business week (dec. , ), http://www.msnbc.msn .com/id/ / (accessed july , ). appendix. project plan: building ha linux platform using cloud computing project manager: project members: object statement: to build a high availability (ha) linux platform to support multiple systems using cloud computing in six months. scope: the project members should identify cloud computing providers, evaluate the costs, and build a linux platform for computer systems, including afghan ils, afghanistan digital libraries website, repository system, japanese inter- library loan website, and digitization management system. resources: project deliverable: january , —july , information technology and libraries | june information technology and libraries | june phase i ■■ to build a stable and reliable linux platform to support multiple web applications. the platform needs to consider reliability and high availability in a cost-effective manner ■■ to install needed libraries for the environment ■■ to migrate ils (koha) to this linux platform ■■ to migrate afghan digital libraries’ website (joomla) to this platform ■■ to migrate japanese interlibrary loan website ■■ to migrate digitization management system phase ii ■■ to research and implement a monitoring tool to monitor all web applications as well as os level tools (e.g. tomcat, mysql) ■■ to configure a cron job to run routine things (e.g., backup ) ■■ to research and implement storage (tb) for digitization and access phase iii ■■ to research and build linux clustering steps: . os installation: debian . platform environment: register dns . install java , tomcat , mysql , etc. . install source control env git . install statistics analysis tool (google analytics) . install monitoring tool: ganglia or nagios . web applications . joomla . koha . monitoring tool . digitization management system . repository system: dspace, fedora, etc. . ha tools/applications note calculation based on the following: ■■ leasing two nodes $ /month: $ x nodes x months = $ /year ■■ a medium-priced server with backup with a life expectancy of years ($ , ): $ , /year ■■ percent of system administrator time for managing the server ($ , annual salary): $ , /year ■■ ignore telecommunication cost, utility cost, and space cost. ■■ ignore software developer’s time because it is equal for both options. appendix. project plan: building ha linux platform using cloud computing (cont.) information technology and libraries | june know its power, and facets can showcase metadata in new interfaces. according to mcguinness, facets perform several functions in an interface: ■■ vocabulary control ■■ site navigation and support ■■ overview provision and expectation setting ■■ browsing support ■■ searching support ■■ disambiguation support these functions offer several potential advantages to the user: the functions use category systems that are coherent and complete, they are predictable, they show previews of where to go next, they show how to return to previous states, they suggest logical alternatives, and they help the user avoid empty result sets as searches are narrowed. disadvantages include the fact that categories of interest must be known in advance, important trends may not be shown, category structures may need to be built by hand, and automated assignment is only partly successful. library catalog records, of course, already supply “categories of interest” and a category structure. information science research has shown benefits to users from faceted search interfaces. but do these benefits hold true for systems as complex as library catalogs? this paper presents an extensive review of both information science and library literature related to faceted browsing. ■■ method to find articles in the library and information science lit- erature related to faceted browsing, the author searched the association for computing machinery (acm) digital library, scopus, and library and information science and technology abstracts (lista) databases. in scopus and the acm digital library, the most successful searches included the following: ■■ (facet* or cluster*) and (usability or user stud*) ■■ facet* and usability in lista, the most successful searches included combining product names such as “aquabrowser” with “usability.” the search “catalog and usability” was also used. the author also searched google and the next generation catalogs for libraries (ngc lib) electronic discussion list in an attempt to find unpublished studies. search terms initially included the concept of “clus- tering”; however, this was quickly shown to be a clearly defined, separate topic. according to hearst, “clustering refers to the grouping of items according to some measure faceted browsing is a common feature of new library catalog interfaces. but to what extent does it improve user performance in searching within today’s library catalog systems? this article reviews the literature for user studies involving faceted browsing and user studies of “next-generation” library catalogs that incorporate faceted browsing. both the results and the methods of these studies are analyzed by asking, what do we cur- rently know about faceted browsing? how can we design better studies of faceted browsing in library catalogs? the article proposes methodological considerations for prac- ticing librarians and provides examples of goals, tasks, and measurements for user studies of faceted browsing in library catalogs. m any libraries are now investigating possible new interfaces to their library catalogs. sometimes called “next-generation library catalogs” or “dis- covery tools,” these new interfaces are often separate from existing integrated library systems. they seek to provide an improved experience for library patrons by offering a more modern look and feel, new features, and the potential to retrieve results from other major library systems such as article databases. one interesting feature these interfaces offer is called “faceted browsing.” hearst defines facets as a “a set of meaningful labels organized in such a way as to reflect the concepts relevant to a domain.” labarre defines fac- ets as representing “the categories, properties, attributes, characteristics, relations, functions or concepts that are central to the set of documents or entities being organized and which are of particular interest to the user group.” faceted browsing offers the user relevant subcategories by which they can see an overview of results, then nar- row their list. in library catalog interfaces, facets usually include authors, subjects, and formats, but may include any field that can be logically created from the marc record (see figure for an example). using facets to structure information is not new to librarians and information scientists. as early as , the classification research group stated a desire to see faceted classification as the basis for all information retrieval. in , ranganathan introduced facet analysis to our profession. librarians like metadata because they jody condit fagan (faganjc@jmu.edu) is content interfaces coordinator, james madison university library, harrisonburg, virginia. jody condit fagan usability studies of faceted browsing: a literature review usability studies of faceted browsing: a literature review | fagan doing so and performed a user study to inform their decision. results: empirical studies of faceted browsing the following summaries present selected empirical research studies that had significant findings related to faceted browsing or inter- esting methods for such studies. it is not an exhaustive list. pratt, hearst, and fagan questioned whether faceted results were better than clustering or relevancy-ranked results. they studied fif- teen breast-cancer patients and families. every subject used three tools: a faceted interface, a tool that clustered the search results, and a tool that ranked the search results according to relevance criteria. the subjects were given three simple queries related to breast cancer (e.g., “what are the ways to prevent breast cancer?”), asked to list answers to these before beginning, and to answer the same queries after using all the tools. in this study, sub- jects completed two timed tasks. first, subjects found as many answers as possible to the question in four minutes. second, the researchers measured the time subjects took to find answers to two specific questions (e.g., “can diet be used in the prevention of breast cancer?”) that related to the original, general query. for the first task, when the subjects used the faceted interface, they found more answers than they did with the other two tools. the mean number of answers found using the faceted interface was . , for the cluster tool it was . , and for the ranking tool it was . . this difference was significant (p< . ). for the second task, the researchers found no significant difference between the tools when comparing time on task. the researchers gave the subjects a user-satisfaction questionnaire at the end of the study. on thirteen of the fourteen quantitative questions, satisfaction scores for the faceted interface were much higher than they were for either the ranking tool or the cluster tool. this difference was statistically significant (p < . ). all fifteen users also affirmed that the faceted interface made sense, was help- ful, was useful, and had clear labels, and said they would use the faceted interface again for another search. yee et al. studied the use of faceted metadata for image searching, and browsing using an interface they developed called flamenco. they collected data from thirty-two participants who were regular users of the internet, searching for information either every day or a few times a week. their subjects performed four tasks (two structured and two unstructured) on each of two interfaces. an example of an unstructured task from their study was “search for images of interest.” an example of a structured task was to gather materials for an art history of similarity . . . typically computed using associations and commonalities among features where features are typically words and phrases.” using library catalog key- words to generate word clouds would be an example of clustering, as opposed to using subject headings to group items. clustering has some advantages according to hearst. it is fully automated, it is easily applied to any text collection, it can reveal unexpected or new trends, and it can clarify or sharpen vague queries. disadvantages to clustering include possible imperfections in the cluster- ing algorithm, similar items not always being grouped into one cluster, a lack of predictability, conflating many dimensions, difficulty labeling groups, and counterintui- tive subhierarchies. in user studies comparing clustering with facets, pratt, hearst, and fagan showed that users find clustering difficult to interpret and prefer a predict- able organization of category hierarchies. ■■ results the author grouped the literature into two categories: user studies of faceted browsing and user studies of library catalog interfaces that include faceted browsing as a feature. generally speaking, the information science literature consisted of empirical studies of interfaces cre- ated by the researchers. in some cases, the researchers’ intent was to create and refine an interface intended for actual use; in others, the researchers created the interface only for the purposes of studying a specific aspect of user behavior. in the library literature, the studies found were generally qualitative usability studies of specific library catalog interface products. libraries had either implemented a new product, or they were thinking about figure . faceted results from jmu’s vufind implementation information technology and libraries | june uddin and janacek asked nineteen users (staff and students at the asian institute of technology) to use a website search engine with both a traditional results list and a faceted results list. tasks were as follows: ( ) look for scholarship information for a masters program, ( ) look for staff recruitment information, and ( ) look for research and associated faculty member information within your interested area. they found that users were faster when using the faceted system, significantly so for two of the three tasks. success in finding relevant results was higher with the faceted system. in the post–study questionnaire, participants rated the faceted system more highly, including significantly higher ratings for flexibil- ity, interest, understanding of information content, and more search results relevancy. participants rated the most useful features to be the capability to switch from one facet to another, preview the result set, combine facets, and navigate via breadcrumbs. capra et al. compared three interfaces in use by the bureau of labor statistics website, using a between-sub- jects study with twenty-eight people and a within-subjects study with twelve people. each set of participants per- formed three kinds of searches: simple lookup, complex lookup, and exploratory. the researchers used an interest- ing strategy to help control the variables in their study: because the bls website is a highly specialized corpus devoted to economic data in the united states orga- nized across very specific time periods (e.g., monthly releases of price or employment data), we decided to include the us as a geographic facet and a month or year as a temporal facet to provide context for all search tasks in our study. thus, the simple lookup tasks were constructed around a single economic facet but also included the spatial and temporal facets to provide context for the searchers. the complex lookup tasks involve additional facets including genre (e.g. press release) and/or region. capra et al. found that users preferred the familiarity afforded by the traditional website interface (hyperlinks + keyword search) but listed the facets on the two experi- mental interfaces as their best features. the researchers concluded, “if there is a predominant model of the infor- mation space, a well designed hierarchical organization might be preferred.” zhang and marchionini analyzed results from fifteen undergraduate and graduate students in a usability study of an interface that used facets to categorize results (relation browser ++). there were three types of tasks: ■■ type : simple look-up task (three tasks such as “check if the movie titled the matrix is in the library movie collection”). ■■ type : data exploration and analysis tasks (six tasks essay on a topic given by the researchers and to complete four related subtasks. the researchers designed the struc- tured task so they knew exactly how many relevant results were in the system. they also gave a satisfaction survey. more participants were able to retrieve all relevant results with the faceted interface than with the baseline interface. during the structured tasks, participants received empty results with the baseline interface more than three times as often as with the faceted interface. the researchers found that participants constructed queries from multiple facets in the unstructured tasks percent of the time and in the structured tasks percent of the time. when given a post–test survey, participants identified the fac- eted interface as easier to use, more flexible, interesting, enjoyable, simple, and easy to browse. they also rated it as slightly more “overwhelming.” when asked to choose between the two, twenty-nine participants chose the fac- eted interface, compared with two who chose the baseline (n = ). thirty-one of the thirty-two participants said the faceted interface helped them learn more, and twenty- eight of them said it would be more useful for their usual tasks. the researchers concluded that even though their faceted interface was much slower than the other, it was strongly preferred by most study participants: “these results indicate that a category-based approach is a suc- cessful way to provide access to image collections.” in a related usability study on the flamenco interface, english et al. compared two image browsing interfaces in a nineteen-participant study. after an initial search, the “matrix view” interface showed a left column with facets, with the images in the result set placed in the main area of the screen. from this intermediary screen, the user could select multiple terms from facets in any order and have the items grouped under any facet. the “singletree” interface listed subcategories of the currently selected term at the top, with query previews underneath. the user could then only drill down to subcategories of the current category, and could not select terms from more than one facet. the researchers found that a majority of participants preferred the “power” and “flexibility” of matrix to the simplicity of singletree. they found it easier to refine and expand searches, shift between searches, and troubleshoot research problems. they did prefer singletree for locating a specific image, but matrix was preferred for browsing and exploring. participants started over only . percent of the time for the matrix compared to . percent for singletree. yet the faceted interface, matrix, was not “better” at everything. for specific image searching, participants found the correct image only . percent of the time in matrix compared to . percent in singletree. also, in matrix, some participants drilled down in the wrong hierarchy with wrong assumptions. one interesting finding was that in both interfaces, more participants chose to begin by browsing ( . percent) than by searching ( . percent). usability studies of faceted browsing: a literature review | fagan of the first two studies: the first study comprised one faculty member, five graduate students, and two under- graduate students; the second comprised two faculty members, four graduate students, and two undergradu- ate students. the third study did not report results related to faceted browsing and is not discussed here. the first study had seven scenarios; the second study had nine. the scenarios were complex: for example, one scenario began, “you want to borrow shakespeare’s play, the tempest, from the library,” but contained the following subtasks as well: . find the tempest. . find multiple editions of this item. . find a recent version. . see if at least one of the editions is available in the library. . what is the call number of the book? . you’d like to print the details of this edition of the book so you can refer to it later. participants found the interface friendly, easy to use, and easy to learn. all the participants reported that fac- eted browsing was useful as a means of narrowing down the result lists, and they considered this tool one of the differentiating features between primo and their library opac or other interfaces. facets were clear, intuitive, and useful to all participants, including opening the “more” section. one specific result from the tests was that “online resources” and “available” limiters were moved from a separate location to the right with all other facets. in a study of aquabrowser by olson, twelve subjects— all graduate students in the humanities—participated in a comparative test in which they looked for additional sources for their dissertation. aquabrowser was created by medialab but is distributed by serials solutions in north america. this study also had three pilot subjects. no relevance judgments were made by the researchers. nine of the twelve subjects found relevant materials by using aquabrowser that they had not found before. olson’s subjects understood facets as a refinement tool (narrowing) and had a clear idea of which facets were useful and not useful for them. they gave overwhelm- ingly positive comments. only two felt the faceted interface was not an improvement. some participants wanted to limit to multiple languages or dates, and a few were confused about the location of facets in multiple places, for example, “music” under both format and topic. a team at yale university, led by bauer, recently conducted two tests on pilot vufind installations: a subject-based presentation of e-books for the cushing/ whitney medical library and a pilot test of vufind using undergraduate students with a sample of , records from the library system. vufind is open-source software developed at villanova university (http://vufind.org). that require users to understand and make sense of the information collection: “in which decade did steven spielberg direct the most movies?”). ■■ type : (one free exploration task: “find five favorite videos without any time constraints”). the tasks assigned for the two interfaces were dif- ferent but comparable. for type tasks, zhang and marchionini found that performance differences between the two interfaces were all statistically significant at the . level. no participants got wrong answers for any but one of the tasks using the faceted interface. with regard to satisfaction, on the exploratory tasks the researchers found statistically significant differences favoring the faceted interface on all three of the satisfaction ques- tions. participants found the faceted interface not as aesthetically appealing nor as intuitive to use as the basic interface. two participants were confused by the constant changing and updating of the faceted interface. the above studies are examples of empirical inves- tigations of experimental interfaces. hearst recently concluded that facets are a “proven technique for sup- porting exploration and discovery” and summarized areas for further research in this area, such as applying facets to large “subject-oriented category systems,” facets on mobile interfaces, adding smart features like “auto- complete” to facets, allowing keyword search terms to affect order of facets, and visualizations of facets. in the following section, user studies of next-generation library catalog interfaces will be presented. results: library literature understandably, most studies by practicing librarians focus on products their libraries are considering for eventual use. these studies all use real library catalog records, usually the entire catalog’s database. in most cases, these studies were not focused on investigating faceted browsing per se, but on the usability of the overall interface. in general, these studies used fewer participants than the information science studies above, followed less rigorous methods, and were not subjected to statistical tests. nevertheless, they provide many insights into the user experience with the extremely complex datasets underneath next-generation library catalog interfaces that feature faceted browsing. in this review article, only results specifically relating to fac- eted browsing will be presented. sadeh described a series of usability studies per- formed at the university of minnesota (um), a primo development partner. primo is the next-generation library catalog product sold by ex libris. the author also received additional information from the usability services lab at um via e-mail. three studies were con- ducted in august , january , and october . eight users from various disciplines participated in each information technology and libraries | june participants. the researchers measured task success, dura- tion, and difficulty, but did not measure user satisfaction. their study consisted of four known-item tasks and six topic-searching tasks. the topic-searching tasks were geared toward the use of facets, for example, “can you show me how would you find the most recently published book about nuclear energy policy in the united states?” all five participants using endeca understood the idea of facets, and three used them. students tried to limit their searches at the outset rather than search and then refine results. an interesting finding was that use of the facets did not directly follow the order in which facets were listed. the most heavily used facet was library of congress classification (lcc), followed closely by topic, and then library, format, author, and genre. results showed a sig- nificantly shorter average task duration for endeca catalog users for most tasks. the researchers noted that none of the students understood that the lcc facet represented call-number ranges, but all of the students understood that these facets “could be used to learn about a topic from dif- ferent aspects—science, medicine, education.” the authors could find no published studies relating to the use of facets in some next-generation library cata- logs, including encore and worldcat local. although the university of washington did publish results of a worldcat local usability study in a recent issue of library technology reports, results from the second round of testing, which included an investigation of facets, were not yet ready. ■■ discussion summary of empirical evidence related to faceted browsing empirical studies in the information science literature support many positive findings related to faceted brows- ing and build a solid case for including facets in search interfaces: ■■ facets are useful for creating navigation structures. ■■ faceted categorization greatly facilitates efficient retrieval in database searching. ■■ facets help avoid dead ends. ■■ users are faster when using a faceted system. ■■ success in finding relevant results is higher with a faceted system. ■■ users find more results with a faceted system. ■■ users also seem to like facets, although they do not always immediately have a positive reaction. ■■ users prefer search results organized into predict- able, multidimensional hierarchies. ■■ participants’ satisfaction is higher with a faceted system. the team drew test questions from user search logs in their current library system. some questions targeted specific problems, such as incomplete spellings and incomplete title information. bauer notes that some problems uncovered in the study may relate to the pecu- liarities of the yale implementation. the medical library study contained eight partici- pants—a mix of medical and nursing students. facets, reported bauer, “worked well in several instances, although some participants did not think they were noticeable on the right side of the page.” the prompt for the faceted task in this study came after the user had done a search: “what if you wanted to look at a particular sub- set, say ‘xxx’ (determine by looking at the facets).” half of the participants used facets, half used “search within” to narrow the topic by adding keywords. sixty-two per- cent of the participants were successful at this task. the undergraduate study asked five participants faced with a results list, “what would you do now if you only wanted to see material written by john adams?” on this task, only one of the five was successful, even though the author’s name was on the screen. bauer noted that in general, “the use of the topic facet to narrow the search was not understood by most participants. . . . even when participants tried to use topic facets the length of the list and extraneous topics rendered them less than useful.” the five undergraduates were also asked, “could you find books in this set of results that are about health and illness in the united states population, or control of com- municable diseases during the era of the depression?” again, only one of the five was successful. bauer notes that “the overly broad search results made this difficult for participants. again, topic facets were difficult to navi- gate and not particularly useful to this search.” bauer’s team noted that when the search was configured to return more hits, “topic facets become a confusingly large set of unrelated items. these imprecise search results, combined with poor topic facet sets, seemed to result in confusion for test participants.” participants were not aware that topics represented subsets, although learning occurred because the “narrow” header was helpful to some par- ticipants. other results found by bauer’s team were that participants were intrigued by facets, navigation tools are needed so that patrons may reorder large sets of topic fac- ets, format and era facets were useful to participants, and call-number facets were not used by anyone. antelman, pace, and lynema studied north carolina state university’s (ncsu) next-generation library catalog, which is driven by software from endeca. their study used ten undergraduate students in a between-subjects design where five used the endeca catalog and five used the library’s traditional catalog. the researchers noted that their participants may have been experienced with the library’s old catalog, as log data shows most ncsu users enter one or two terms, which was not true of study usability studies of faceted browsing: a literature review | fagan one product’s faceted system for a library catalog does not substitute for another, the size and scope of local collections may greatly affect results, and cataloging practices and metadata will affect results. still, it is important for practic- ing librarians to determine if new features such as facets truly improve the user’s experience. methodological best practices after reading numerous empirical research studies (some of which critique their own methods) and library case studies, some suggestions for designing better studies of facets in library catalogs emerged. designing the study ■■ consider reusing protocols from previous studies. this provides not only a tested method but also a possible point of comparison. ■■ define clear goals for each study and focus on spe- cific research questions. it’s tempting to just throw the user into the interface and see what happens, but this makes it difficult, if not impossible, to analyze the results in a useful way. for example, one of zhang and marchionini’s hypotheses specifically describes what rich interaction would look like: “typing in key- words and clicking visual bars to filter results would be used frequently and interchangeably by the users to finish complex search tasks, especially when large numbers of results are returned.” ■■ develop the study for one type of user. olson’s focus on graduate students in the dissertation process allowed the researchers to control for variables such as interest of and knowledge about the subject. ■■ pilot test the study with a student worker or col- league to iron out potential wrinkles. ■■ let users explore the system for a short time and pos- sibly complete one highly structured task to help the user become used to the test environment, interface, and facilitator. unless you are truly interested in the very first experience users have with a system, the first use of a system is an artificial case. designing tasks ■■ make sure user performance on each task is measur- able. will you measure the time spent on a task? if “success” is important, define what that would look like. for example, english et al. defined success for one of their tasks as when “the participant indicated (within the allotted time) that he/she had reached an appropriate set of images/specific image in the collection.” ■■ establish benchmarks for comparison. one can test for significant differences between interfaces, one can test for differences between research subjects and an expert user, and one can simply measure against ■■ users are more confident with a faceted system. ■■ users may prefer the familiarity afforded by tra- ditional website interface (hyperlinks + keyword search). ■■ initial reactions to the faceted interface may be cau- tious, seeing it as different or unfamiliar. users interact with specific characteristics of faceted interfaces, and they go beyond just one click with facets when it is permitted. english et al. found that percent of their participants expanded facets by removing a term, and that facets were used more than “keyword search within”: . percent versus percent. yee et al. found that participants construct queries from multiple facets percent of the time in unstructured tasks; in structured tasks they do so percent of the time. the above studies did not use library catalogs; in most cases they used an experimental interface with record sets that were much smaller and less complicated than in a complete library collection. domains included websites, information from one website, image collections, video collections, and a journal article collection. summary of practical user studies related to faceted browsing this review also included studies from practicing librar- ians at live library implementations. these studies generally had smaller numbers of users, were more likely to focus on the entire interface rather than a few features, and chose more widely divergent methods. studies were usually linked to a specific product, and results varied widely between systems and studies. for this reason it is difficult to assemble a bulleted summary as with the previous section. the variety of results from these studies indicate that when faceted browsing is applied to a real- life situation, implementation details can greatly affect user performance and user preference. some, like labarre, are skeptical about whether fac- ets are appropriate for library information. descriptions of library materials, says labarre, include analyses of intellectual content that go beyond the descriptive terms assigned to commercial items such as a laptop: now is the time to question the assumptions that are embedded in these commercial systems that were primarily designed to provide access to concrete items through descriptions in order to enhance profit. it is clear that an evaluation of commercial interfaces or experimental interfaces does not substitute for an opac evaluation. yet it is a challenge for libraries to find expertise and resources to conduct user studies. the systems they want to test are large and complex. collaborating with other libraries has its own challenges: an evaluation of information technology and libraries | june groups of participants, each of which tests a dif- ferent system. ■❏ a within-subjects design has one group of par- ticipants test both systems. it is hoped that if libraries use the suggestions above when designing future experiments, results across studies will be more comparable and useful. designing user studies of faceted browsing after examining both empirical research studies and case studies by practicing librarians, a key difference seems to be the specificity of research questions and design- ing tasks and measurements to test specific hypotheses. while describing a full user-study protocol for investi- gating faceted browsing in a library catalog is beyond the scope of this article, reviewing the literature and the study methods it describes provided insights into how hypotheses, tasks, and measurements could be written to provide more reliable and comparable evidence related to faceted browsing in library catalog systems. for example, one research question could surround the format facet: “compared with our current interface, does our new faceted interface improve the user’s ability to find different formats of materials?” hypotheses could include the following: . users will be more accurate when identifying the formats of items from their result set when using the faceted interface than when using the traditional interface. . users will be able to identify formats of items more quickly with the faceted interface than with the tradi- tional interface. looking at these hypotheses, here is a prompt and some example tasks the participants would be asked to perform: “we will be asking you to find a variety of for- mats of materials. when we say formats of materials, we mean books, journal articles, videos, etc.” ■■ task : please use interface a to search on “interper- sonal communication.” look at your results set. please list as many different formats of material as you can. ■■ task : how many items of each format are there? ■■ task : please use interface b to search on “family communication.” what formats of materials do you see in your results set? ■■ task : how many items of each format are there?” we would choose the topics “interpersonal com- munication” and “family communication” because our local catalog has many material types for these topics and because these topics would be understood by most of our students. we would choose different topics to expectations or against previous iterations of the same study. for example, “ percent of users completed the task within five minutes.” zhang and marchionini measured error rates, another possible benchmark. ■■ consider looking at your existing opac logs for zero- results searches or other issues that might inspire interesting questions. ■■ target tasks to avoid distracters. for example, if your catalog has a glut of government documents, consider running the test with a limit set to exclude them unless you are specifically interested in their impact. for example, capra et al. decided to include the united states as a geographic facet and a month or year as a temporal facet to provide context for all search tasks in their study. ■■ for some tasks, give the subjects simple queries (e.g., “what are the ways to prevent breast cancer?”) as opposed to asking the subjects to come up with their own topic. this can help control for the potential challenges of formulating one’s own research ques- tion on the spot. as librarians know, formulating a good research question is its own challenge. ■■ if you are using any timed tasks, consider how the nature of your tasks could affect the result. for example, pratt, hearst, and fagan noted that the time that it took subjects to read and understand abstracts most heavily influenced the time for them to find an answer. english et al. found that the system’s pro- cessing time influenced their results. ■■ consider the implications of your local implementa- tion carefully when designing your study. at yale, the team chose to point their vufind instance at just , of their records, drew questions from prob- lems users were having (as shown in log files), and targeted questions to these problems. who to study? ■■ try to study a larger set of users. it is better to create a short test with many users than a long test with a few users. nielsen suggests that twenty users is suf- ficient. consider collaborating with another library if necessary. ■■ if you test a small number, such as the typical four to eight users for a usability test, be sure you emphasize that your results are not generalizable. ■■ use subjects who are already interested in the subject domain: for example, pratt, hearst, and fagan used breast cancer patients, and olson used graduate students currently writing their dissertations. ■■ consider focusing on advanced or scholarly users. la barre suggests that undergraduates may be over- studied. ■■ for comparative studies, consider having both between-subjects and within-subjects designs. ■❏ a between-subjects design involves creating two usability studies of faceted browsing: a literature review | fagan these experimental studies. previous case-study inves- tigations of library catalog interfaces with facets have proven inconclusive. by choosing more specific research questions, tasks, and measurements for user studies, libraries may be able to design more objective studies and compare results more effectively. references . marti a. hearst, “clustering versus faceted categories for information exploration,” communications of the acm , no. ( ): . . kathryn la barre, “faceted navigation and browsing fea- tures in new opacs: robust support for scholarly information seeking?” knowledge organization , no. ( ): . . vanda broughton, “the need for faceted classification as the basis of all methods of information retrieval,” aslib proceed- ings , no. / ( ): – . . s. r. ranganathan, colon classification basic classification, th ed. (new york: asia, ). . deborah l. mcguinness, “ontologies come of age,” in spinning the semantic web: bringing the world wide web to its full potential, ed. dieter fensel et al. (cambridge, mass.: mit pr., ): – . . hearst, “clustering versus faceted categories,” . . ibid., . . ibid., . . ibid.. . . wanda pratt, marti a. hearst, and lawrence m. fagan, “a knowledge-based approach to organizing retrieved docu- ments,” proceedings of the sixteenth national conference on artifi- cial intelligence, july – , , orlando, florida (menlo park, calif.: aaai pr., ): – . . ibid. . ibid., . . ka-ping yee et al., “faceted metadata for image search and browsing,” , http://flamenco.berkeley.edu/papers/ flamenco-chi .pdf (accessed oct. , ). . ibid., . . ibid., . . ibid. . ibid., . . jennifer english et al., “flexible search and navigation,” , http://flamenco.berkeley.edu/papers/flamenco .pdf (accessed apr. , ). . ibid., . . ibid., . . ibid., . . mohammed nasir uddin and paul janecek, “performance and usability testing of multidimensional taxonomy in web site search and navigation,” performance measurement and met- rics , no. ( ): – . . ibid., . . robert capra et al., “effects of structure and interaction style on distinct search tasks,” proceedings of the th acm-ieee-cs joint conference on digital libraries (new york: acm, ): – . . ibid., . . ibid., . help minimize learning effects. to further address this, we would plan to have half our users start first with the traditional interface and half to start first with the faceted interface. this way we can test for differences resulting from learning. the above tasks would allow us to measure several pieces of evidence to support or reject our hypotheses. for tasks and , we would measure the number of formats correctly identified by users compared with the number found by an expert searcher. for tasks and , we would compare the number of items correctly identified with the total items found in each category by an expert searcher. we could also time the user to determine which interface helped them work more quickly. in addition to measuring the number of formats identified and the number of items identified in each format, we would be able to measure the time it takes users to identify the number of formats and the number of items in each format. to measure user satisfaction, we would ask participants to complete the system usability scale (sus) after each interface and, at the very end of the study, complete a questionnaire com- paring the two interfaces. even just selecting the format facet, we would have plenty to investigate. other hypotheses and tasks could be developed for other facet types, such as time period or publication date, or facets related to the responsible par- ties, such as author or director: hypothesis: users can find more materials written in a certain time period using the faceted interface. task: find ten items of any type (books, journals, mov- ies) written in the s that you think would have information about television advertising. hypothesis: users can find movies directed by a spe- cific person more quickly using the faceted interface. task: in the next two minutes, find as many movies as you can that were directed by orson welles. for the first task above, an expert searcher could complete the same task, and their time could be used as a point of comparison. for the second, the total number of movies in the library catalog that were directed by welles is an objective quantity. in both cases, one could compare the user’s performance on the two interfaces. ■■ conclusion reviewing user studies about faceted browsing revealed empirical evidence that faceted browsing improves user performance. yet this evidence does not necessarily point directly to user success in faceted library catalogs, which have much more complex databases than those used in information technology and libraries | june . uddin and janecek, “performance and usability testing”; zhang and marchionini, evaluation and evolution; hao chen and susan dumais, bringing order to the web: automatically categoriz- ing search results (new york: acm, ): – . . uddin and janecek, “performance and usability testing.” . ibid.; pratt, hearst, and fagan, “a knowledge-based approach”; hsinchun chen et al., “internet browsing and searching: user evaluations of category map and concept space techniques,” journal of the american society for information science , no. ( ): – . . vanda broughton, “the need for faceted classification as the basis of all methods of information retrieval,” aslib proceedings , no. / ( ): – ; pratt, hearst, and fagan, “a knowledge-based approach,” – .; chen et al., “internet browsing and searching,” – ; yee et al., “faceted metadata for image search and browsing”; english et al., “flexible search and navigation using faceted metadata.” . uddin and janecek, “performance and usability testing”; zhang and marchionini, evaluation and evolution; hideo joho and joemon m. jose, slicing and dicing the information space using local contexts (new york: acm, ): – .; yee et al., “faceted metadata for image search and browsing.” . yee et al., “faceted metadata for image search and brows- ing”; chen and dumais, bringing order to the web. . capra et al., “effects of structure and interaction style.” . yee et al., “faceted metadata for image search and brows- ing”; capra et al., “effects of structure and interaction style”; zhang and marchionini, evaluation and evolution. . english et al., “flexible search and navigation,” . . yee et al., “faceted metadata for image search and brows- ing,” . . la barre, “faceted navigation and browsing,” . . zhang and marchionini, evaluation and evolution, . . english et al., “flexible search and navigation.” . ibid., . . zhang and marchionini, evaluation and evolution. . capra et al., “effects of structure and interaction style.” . pratt, hearst, and fagan, “a knowledge-based approach.” . english et al., “flexible search and navigation.” . bauer, “yale university library vufind test—under- graduates.” . jakob nielsen, “quantitative studies: how many users to test?” online posting, alertbox, june , http://www.useit .com/alertbox/quantitative_testing.html (accessed apr. , ). . pratt, hearst, and fagan, “a knowledge-based approach.” . tod a. olson used graduate students currently writing their dissertations. olson, “utility of a faceted catalog for schol- arly research,” library hi tech , no. ( ): – . . la barre, “faceted navigation and browsing.” . capra et al., “effects of structure and interaction style.” . junliang zhang and gary marchionini, evaluation and evolution of a browse and search interface: relation browser++ (atlanta, ga.: digital government society of north america, ): – . . ibid., . . marti a. hearst, “uis for faceted navigation: recent advances and remaining open problems,” , http://people. ischool.berkeley.edu/~hearst/papers/hcir .pdf (accessed apr. , ). . tamar sadeh, “user experience in the library: a case study,” new library world , no. / (jan. ): – . . ibid., . . jerilyn veldof, e-mail from university of minnesota usability services lab, . . tod a. olson, “utility of a faceted catalog for scholarly research,” library hi tech , no. ( ): – . . ibid., . . kathleen bauer, “yale university library vufind test— undergraduates,” may , , http://www.library.yale.edu/ usability/studies/summary_undergraduate.doc (accessed apr. , ); kathleen bauer and alice peterson-hart, “usability test of vufind as a subject-based display of ebooks,” aug. , , http://www.library.yale.edu/usability/studies/summary _medical.doc (accessed apr. , ). . bauer and peterson-hart, “usability test of vufind as a subject-based display of ebooks,” . . ibid., . . ibid., . . ibid. . ibid., . . ibid. . ibid., . . ibid., . . kristin antelman, andrew k. pace, and emily lynema, “toward a twenty-first century library catalog,” information technology & libraries , no. ( ): – . . ibid., . . ibid., . . ibid., . . ibid., . . jennifer l. ward, steve shadle, and pam mofield, “user experience, feedback, and testing,” library technology reports , no. (aug. ): . . english et al., “flexible search and navigation.” . peter ingwersen and irene wormell, “ranganathan in the perspective of advanced information retrieval,” libri ( ): – ; winfried godert, “facet classification in online retrieval,” international classification , no. ( ): – .; w. godert, “klassificationssysteme und online-katalog [classifica- tion systems and the online catalogue],” zeitschrift für biblio- thekswesen und bibliographie , no. ( ): – . . yee et al., “faceted metadata for image search and brows- ing”; english et al., “flexible search and navigation.” editorial | truitt marc truitt marc truitt (marc.truitt@ualberta.ca) is associate university librarian, bibliographic and information technology services, university of alberta libraries, edmonton, alberta, canada, and editor of ital. marc truitt editorial: and now for something (completely) different t he issue of ital you hold in your hands—be that issue physical or virtual; we won’t even go into the question of your hands!—represents something new for us. for a number of years, ex libris (and previ- ously, endeavor information systems) has generously sponsored the lita/ex libris (née lita/endeavor) student writing award competition. the competition seeks manuscript submissions from enrolled lis students in the areas of ital’s publishing interests; a lita committee on which the editor of ital serves as an ex-officio member evaluates the entries and names a winner. traditionally, the winning essay has appeared in the pages of ital. in recent years, perhaps mirroring the waning interest in publication in traditional peer- reviewed venues, the number of entrants in the competi- tion has declined. in , for instance, there were but nine submissions, and to get those, we had to extend the deadline six weeks from the end of february to mid- april. in previous years, as i understand it, there often were even fewer. this year, without moving the goalposts, we had— hold onto your hats!—twenty-seven entries. of these, the review committee identified six finalists for discussion. the turnout was so good, in fact, that with the agreement of the committee, we at ital proposed to publish not only the winning paper but the other finalist entries as well. we hope that you will find them as stimulating as have we. even more importantly, we hope that by pub- lishing such a large group of papers representing ’s best in technology-focused lis work, we will encourage similarly large numbers of quality submissions in the years to come. i would like to offer sincere thanks to my university of alberta colleague sandra shores, who as guest editor for this issue worked tirelessly over the past few months to shepherd quality student papers into substantial and interesting contributions to the literature. she and managing editor judith carter—who guest-edited our recent discovery issue—have both done fabulous jobs with their respective ital special issues. bravo! n ex libris’ sponsorship in one of those ironic twists that one more customarily associates with movie plots than with real life, the lita/ex libris student writing award recently almost lost its spon- sor. at very nearly the same time that sandra was complet- ing the preparation of the manuscripts for submission to ala production services (where they are copyedited and typeset), we learned that ex libris had notified lita that it had “decided to cease sponsoring” the student writing award. a brief round of e-mails among principals at lita, ex libris, and ital ensued, with the outcome being that carl grant, president of ex libris north america, gra- ciously agreed to continue sponsorship for another year and reevaluate underwriting the award for the future. we at ital and i personally are grateful. carl’s message about the sponsorship raises some interesting issues on which i think we should reflect. his first point goes like this: it simply is not realistic for libraries to continue to believe that vendors have cash to fund these things at the same levels when libraries don’t have cash to buy things (or want to delay purchases or buy the product for greatly reduced amounts) from those same vendors. please understand the two are tied together. point taken and conceded. money is tight. carl’s argu- ment, i think, speaks as well to a larger, implied question. libraries and library vendors share highly synergistic and, in recent years, increasingly antagonistic relation- ships. library vendors—and i think library system ven- dors in particular—come in for much vitriol and precious little appreciation from those of us on the customer side. we all think they charge too much (and by implication, must also make too much), that their support and service are frequently unresponsive to our needs, and that their systems are overly large, cumbersome, and usually don’t do things the way we want them done. at the same time, we forget that they are catering to the needs and whims of a small, highly specialized market that is characterized by numerous demands, a high degree of complexity, and whose members—“standards” notwithstanding—rarely perform the same task the same way across institutions. we expect very individualized service and support, but at the same time are penny-pinching misers in our ability and willingness to pay for these services. we are beggars, yet we insist on our right to be choosers. finally, at least for those of us of a certain generation—and yep, i count myself among its members—we chose librarianship for very specific reasons, which often means we are more than a little uneasy with concepts of “profit” and “bottom line” as applied to our world. we fail to understand the open-source dictum that “free as in kittens and not as in beer” means that we will have to pay someone for these services—it’s only a question of whom we will pay. carl continues, making another point: i do appreciate that you’re trying to provide us more recognition as part of this. frankly, that was another consideration in our thought of dropping it—we just didn’t feel like we were getting much for it. marc truitt (marc.truitt@ualberta.ca) is associate university librarian, bibliographic and information technology services, university of alberta libraries, edmonton, alberta, canada, and editor of ital. information technology and libraries | march i’ve said before and i’ll say again, i’ve never, in all my years in this business had a single librarian say to me that because we sponsored this or that, it was even a consideration in their decision to buy something from us. not once, ever. companies like ours live on sales and service income. i want to encourage you to help make librarians aware that if they do appreciate when we do these things, it sure would be nice if they’d let us know in some real tangible ways that show that is true. . . . good will does not pay bills or salaries unless that good will translates into purchases of products and services (and please note, i’m not just speaking for ex libris here, i’m saying this for all vendors). and here is where carl’s and my views may begin to diverge. let’s start by drawing a distinction between vendor tchotchkes and vendor sponsorship. in fairness, carl didn’t say anything about tchotchkes, so why am i? i do so because i think that we need to bear in mind that there are multiple ways vendors seek to advertise themselves and their services to us, and geegaws are one such. trinkets are nice—i have yet to find a better gel pen than the ones given out at iug (would that i could get more!)—but other than reminding me of a vendor’s name, they serve little useful purpose. the latter, vendor sponsorship, is something very different, very special, and not readily totaled on the bottom line. carl is quite right that sponsorship of the student writing award will not in and of itself cause me to buy aleph, primo, or sfx (oh right, i have that last one already!). these are products whose purchase is the result of lengthy and complex reviews that include highly detailed and painstaking needs analysis, specifications, rfps, site visits, demonstrations, and so on. due diligence to our parent institutions and obligations to our users require that we search for a balance among best-of-breed solutions, top-notch support, and fair pricing. those things aren’t related to sponsorship. what is related to sponsorship, though, is a sense of shared values and interests. of “doing the right thing.” i may or may not buy carl’s products because of the con- siderations above (and yes, ex libris fields very strong contenders in all areas of library automation); i definitely will, though, be more likely to think favorably of ex libris as a company that has similar—though not necessarily identical—values to mine, if it is obvious that it encour- ages and materially supports professional activities that i think are important. support for professional growth and scholarly publication in our field are two such values. i’m sure we can all name examples of this sort of behavior: in addition to support of the student writing award, ex libris’ long-standing prominence in the national information standards organization (niso) comes to mind. so too does the founding and ongoing support by innovative interfaces and the library consulting firm r for the taiga forum (http://www.taigaforum.org/), a group of academic associate university librarians. to the degree that i believe ex libris or another firm shares my values by supporting such activities—that it “does the right thing”—i will be just a bit more inclined to think positively of it when i’m casting about for solutions to a technology or other need faced by my institution. i will think of that firm as kin, if you will. with that, i will end this by again thanking carl and ex libris—because we don’t say thank you often enough!—for their generous support of the lita/ex libris student writing award. i hope that it will continue for a long time to come. that support is something about which i do care deeply. if you feel similarly—be it about the student writing award, niso, taiga, or whatever—i urge you to say so by sending an appropriate e-mail to your vendor’s representative or by simply saying thanks in person to the company’s head honcho on the ala exhibit floor. and the next time you are neck-deep in seemingly identical vendor quotations and need a way to figure out how to decide between them, remember the importance of shared values. n dan marmion longtime lita members and ital readers in particu- lar will recognize the name of dan marmion, editor of this journal from through . many current and recent members of the ital editorial board—including managing editor judith carter, webmaster andy boze, board member mark dehmlow, and i—can trace our involvement with ital to dan’s enthusiastic period of stewardship as editor. in addition to his leadership of ital, dan has been a mentor, colleague, boss, and friend. his service philoso- phy is best summarized in the words of a simple epigram that for many years has graced the wall behind the desk in his office: “it’s all about access!!” because of health issues, and in order to devote more time to his wife diana, daughter jennifer, and grand- daughter madelyn, dan recently decided to retire from his position as associate director for information systems and digital access at the university of notre dame hesburgh libraries. he also will pursue his personal interests, which include organizing and listening to his extensive collection of jazz recordings, listening to books on cd, and following the exploits of his favorite sports teams, the football irish of notre dame, the indianapolis colts, and the new york yankees. we want to express our deep gratitude for all he has given to the profession, to lita, to ital, and to each of us personally over many years. we wish him all the best as he embarks on this new phase of his life. information technology and libraries | march sandra shores is [tk] sandra shores editorial board thoughts: issue introduction to student essays t he papers in this special issue, although covering diverse topics, have in common their authorship by people currently or recently engaged in gradu- ate library studies. it has been many years since i was a library science student—twenty-five in fact. i remember remarking to a future colleague at the time that i found the interview for my first professional job easy, not because the interviewers failed to ask challenging questions, but because i had just graduated. i was passionate about my chosen profession, and my mind was filled from my time at library school with big ideas and the latest theories, techniques, and knowledge of our discipline. while i could enthusiastically respond to anything the interviewers asked, my colleague remarked she had been in her job so long that she felt she had lost her sense of the big questions. the busyness of her daily work life drew her focus away from contemplation of our purpose, principles, and values as librarians. i now feel at a similar point in my career as this colleague did twenty-five years ago, and for that reason i have been delighted to work with these student authors to help see their papers through to publication. the six papers represent the strongest work from a wide selection that students submitted to the lita/ ex libris student writing award competition. this year’s winner is michael silver, who looks for- ward to graduating in the spring from the mlis program at the university of alberta. silver entered the program with a strong library technology foundation, having pro- vided it services to a regional library system for about ten years. he notes that “the ‘accidental systems librarian’ position is probably the norm in many small and medium sized libraries. as a result, there are a number of practices that libraries should adopt from the it world that many library staff have never been exposed to.” his paper, which details the implementation of an open-source mon- itoring system to ensure the availability of library systems and services, is a fine example of the blending of best practices from two professions. indeed, many of us who work in it in libraries have a library background and still have a great deal to learn from it professionals. silver is contemplating a phd program or else a return to a library systems position when he graduates. either way, the pro- fession will benefit from his thoughtful, well-researched, and useful contributions to our field. todd vandenbark’s paper on library web design for persons with disabilities follows, providing a highly prac- tical but also very readable guide for webmasters and others. vandenbark graduated last spring with a mas- ters degree from the school of library and information science at indiana university and is already working as a web services librarian at the eccles health sciences library at the university of utah. like mr. silver, he entered the program with a number of years’ work experience in the it field, and his paper reflects the depth of his technical knowledge. vandenbark notes, however, that he has found “the enthusiasm and collegiality among library technology professionals to be a welcome change from other employment experiences,” a gratifying com- ment for readers of this journal. ilana tolkoff tackles the challenging concept of global interoperability in cataloguing. she was fascinated that a single database, oclc, has holdings from libraries all over the world. this is also such a recent phenom- enon that our current cataloging standards still do not accommodate such global participation. i was inter- ested to see what librarians were doing to reconcile this variety of languages, scripts, cultures, and indepen- dently developed cataloging standards. tolkoff also graduated this past spring and is hoping to find a position within a music library. marijke visser addresses the overwhelming question of how to organize and expose internet resources, looking at tagging and the social web as a solution. coming from a teaching background, visser has long been interested in literacy and life-long learning. she is concerned about “the amount of information found only online and what it means when people are unable . . . to find the best resources, the best article, the right website that answers a question or solves a critical problem.” she is excited by “the potential for creativity made possible by technol- ogy” and by the way librarians incorporate “collaborative tools and interactive applications into library service.” visser looks forward to graduating in may. mary kurtz examines the use of the dublin core metadata schema within dspace institutional repositor- ies. as a volunteer, she used dspace to archive historical photographs and was responsible for classifying them using dublin core. she enjoyed exploring how other institutions use the same tools and would love to delve further into digital archives, “how they’re used, how they’re organized, who uses them and why.” kurtz graduated in the summer and is looking for the right job for her interests and talents in a location that suits herself and her family. finally, lauren mandel wraps up the issue exploring the use of a geographic information system to under- stand how patrons use library spaces. mandel has been an enthusiastic patron of libraries since she was a small child visiting her local county and city public libraries. she is currently a doctoral candidate at florida state university and sees an academic future for herself. mandel expresses infectious optimism about technology in libraries: sandra shores (sandra.shores@ualberta.ca) is guest editor of this issue and operations manager, information technology servi- ces, university of alberta libraries, edmonton, alberta, canada. editorial board thoughts | shores looking ahead, it seems clear that the pace of change in today’s environment will only continue to accelerate; thus the need for us to quickly form and dissolve key sponsorships and partnerships that will result in the suc- cessful fostering and implementation of new ideas, the currency of a vibrant profession. the next challenge is to realize that many of the key sponsorship and partnerships that need to be formed are not just with traditional organizations in this profession. tomorrow’s sponsorships and partnership will be with those organizations that will benefit from the expertise of libraries and their suppliers while in return helping to develop or provide the new funding opportunities and means and places for disseminating access to their expertise and resources. likely organizations would be those in the fields of education, publishing, content cre- ation and management, and social and community web- based software. to summarize, we at ex libris believe in sponsor- ships and partnerships. we believe they’re important and should be used in advancing our profession and organizations. from long experience we also have learned there are right ways and wrong ways to implement these tools, and i’ve shared thoughts on how to make them work for all the parties involved. again, i thank marc for his receptiveness to this discussion and my even deeper appreciation for trying to address the issues. it’s serves as an excellent example of what i discussed above. people forget, but paper, the scroll, the codex, and later the book were all major technological leaps, not to mention the printing press and moveable type. . . . there is so much potential for using technology to equalize access to information, regardless of how much money you have, what language you speak, or where you live. big ideas, enthusiasm, and hope for the profession, in addition to practical technology-focused information await the reader. enjoy the issue, and congratulations to the winner and all the finalists! note . all quotations are taken with permission from private e-mail correspondence. a partnership for creating successful partnerships continued from page from our readers | eden bradford lee edenfrom our readers the new user environment: the end of technical services? editor’s note: “from our readers” is an occasional feature high- lighting ital readers’ letters and commentaries on timely issues. technical services: an obsolete term used to describe the largest component of most library staffs in the twentieth century. that component of the staff was entirely devoted to arcane and mysterious processes involved in selecting, acquiring, cataloging, pro- cessing, and otherwise making available to library users physical material containing information con- tent pieces (incops). the processes were compli- cated, expensive, and time-consuming, and generally served to severely limit direct service to users both by producing records that were difficult to under- stand and interpret, even by other library staff, and by consuming from – percent of the library’s financial and personnel resources. in the twenty-first century, the advent of new forms of publication and new techniques for providing universal records and universal access to information content made the organizational structure obsolete. that change in organizational structure, more than any other single factor, is generally credited as being responsible for the dramatic improvement in the quality of library service that has occurred in the first decade of the twenty-first century. t here are many who would say that i was the one who wrote this quotation. i didn’t, and it is, in fact, more than twenty-five years old! while i was beginning to research and prepare for this article, i began as most users today start their search for information: i started with google. granted, i rarely go beyond the first page of results (as most user surveys indicate), but the paucity of links made me click to the next screen. there, at number , was a scanned article. jackpot! i thought as i started perusing the contents of this resource online, thinking to myself how the future had changed so dramatically since , with the emergence of the internet and the laptop, all of the new information formats, and the digitization of information. ahh, the power of full text! after reading through the table of contents, introduction, and the first chapter, i noticed that some of the pages were missing. mmmm, obviously some very shoddy scanning on the part of google. but no, i finally realized that only part of this special issue was available on google. obviously, i missed the statement at the bottom of the front scan of the book: “this is a preview. the total pages displayed will be limited. learn more.” and thus the issues regarding copy- right reared their ugly head. when discussing the new user environment, there are many demands facing libraries today. in a report by martha bates, citing the principle of least effort first attributed to philologist george zipf and quoted in the calhoun report to the library of congress, she states: people do not just use information that is easy to find; they even use information that they know to be of poor quality and less reliable—so long as it requires little effort to find—rather than using information they know to be of high quality and reliable, though harder to find . . . despite heroic efforts on the part of librarians, students seldom have sufficiently sustained exposure to and practice with library skills to reach the point where they feel real ease with and mastery of library information systems. according to the final report of bibliographic services task force of the university of california libraries, users expect the following: ■■ one system or search to cover a wide information universe (e.g., google or amazon) ■■ enriched metadata (e.g., onix, tables of contents, and cover art) ■■ full-text availability ■■ to move easily and seamlessly from a citation about an item to the item itself—discovery alone is not enough ■■ systems to provide a lot of intelligent assistance ■❏ correction of obvious spelling errors ■❏ results sorting in order of relevance to their queries ■❏ help in navigating large retrievals through logi- cal subsetting or topical maps or hierarchies ■❏ help in selecting the best source through rel- evance ranking or added commentary from peers and experts or “others who used this also used that” tools ■❏ customization and personalization services ■■ authenticated single sign-on ■■ security and privacy ■■ communication and collaboration ■■ multiple formats available: e-books, mpeg, jpeg, rss and other push technologies, along with tradi- tional, tangible formats ■■ direct links to e-mail, instant messaging, and sharing ■■ access to online virtual communities ■■ access to what the library has to offer without actu- ally having to visit the library bradford lee eden (eden@library.ucsb.edu) is associate uni- versity librarian for technical services & scholarly communica- tion, university of california, santa barbara. information technology and libraries | june what is there in this new user environment for those who work in technical services? as indicated in the open- ing quote, would a dramatic improvement in library services occur if technical services were removed from the organizational structure? even in , the huge financial investment that libraries made in the organization and description of information, inventory, workflows, and personnel was recognized; today, that investment comes under intense scrutiny as libraries realize that we no longer have a monopoly on information access, and to survive we need to move forward more aggressively into the digital environment than ever before. as marcum stated in her now-famous article, ■■ if the commonly available books and journals are accessible online, should we consider the search engines the primary means of access to them? ■■ massive digitization radically changes the nature of local libraries. does it make sense to devote local efforts to the cataloging of unique materials only rather than the regular books and journals? ■■ we have introduced our cataloging rules and the marc format to libraries all over the world. how do we make massive changes without creating chaos? ■■ and finally, a more specific question: should we proceed with aacr in light of a much-changed environment? there are larger internal issues to consider here as well. the budget situation in libraries requires the application of business models to workflows that have normally not been questioned nor challenged. karen calhoun discusses this topic in a number of her contribu- tions to the literature: when catalog librarians identify what they contribute to their communities with their methods (the catalog- ing rules, etc.) and with the product they provide (the catalog), they face the danger of “marketing myopia.” marketing myopia is a term used in the business litera- ture to describe a nearsighted view that focuses on the products and services that a firm provides, rather than the needs those products and services are intended to address. for understanding the implementation issues associ- ated with the leadership strategy, it is important to be clear about what is meant by the “excess capacity” of catalogs. most catalogers would deny there is excess capacity in today’s cataloging departments, and they are correct. library materials continue to flood into acquisitions and cataloging departments and staff can barely keep up. yet the key problem of today’s online catalog is the effect of declining demand. in healthy businesses, the demand for a product and the capacity to produce it are in balance. research libraries invest huge sums in the infrastructure that produces their local catalogs, but search engines are students and scholars’ favorite place to begin a search. more users bypass catalogs for search engines, but research librar- ies’ investment in catalogs—and in the collections they describe—does not reflect the shift in user demand. i have discussed this exact problem in recent articles and technical reports as well. there have to be better, more efficient ways for libraries to organize and describe information not based on the status quo of redundant “localizing” of bibliographic records. a good analogy would be the current price of gas and the looming trans- portation crisis. for many years, americans have had the luxury of being able to purchase just about any type of car, truck, suv, hummer, etc., that they wanted on the basis of their own preferences, personalities, and incomes, not on the size of the gas tank or on the mileage per gallon. why not buy a mercedes over a kia? but with gas prices now well above the average person’s ability to consistently fill their gas tank without mortgaging their future, the market demands that people find alternative solutions in order to survive. this has meant moving away from the status quo of personal choice and selec- tion toward a more economic and sustainable model of informed fuel-efficiency transportation, so much so that public transportation is now inundated with more users than it can handle, and consumers have all but abandoned the truck and suv markets. libraries have long worked in the mercedes arena, providing features such as authority control, subject classification, and redundant localizing of bibliographic records that were essential when libraries held the monopoly on informa- tion access but are no longer cost-efficient—nor even sane—strategies in the current information marketplace. users are not accessing the opac anymore; well-known studies indicate that more than percent of informa- tion seekers begin their search on a web search engine. libraries are investing huge resources in staffing and priorities fiddling with marc bibliographic records in a time when they are struggling to survive and adapt from a monopoly environment to being just one of many players in the new information marketplace. budgets are stagnant, staffing is at an all-time low, new information formats continue to appear and require attention, and users are no longer patient nor comfortable working with our clunky opacs. why do libraries continue to support an infrastructure of buying and offering the same books, cds, dvds, journals, etc., at every library, when the new information environment offers libraries the opportu- nity to showcase and present their unique information resources and one-of-a-kind collections to the world? special collections materials held by every major research and public library in the world can now be digitized, and from our readers | eden sparse library resources need to be adjusted to compete and offer these unique collections and their services to our users and the world. the october issue of computers in libraries is devoted solely to articles related to the enhancement, usability, appropriateness, and demise of the library opac. interesting articles include “fac-back-opac: an open source solution interface to your library system,” “dreaming of a better ils,” “plug your users into library resources with opensearch plug-ins,” delivering what people need, when and where they need it,” “the birth of a new generation of library interfaces,” and “will the ils soon be as obsolete as the card catalog?” an especially interesting quote is given by cervone, then assistant university librarian for information technology at northwestern university: what i’d like to see is for the catalog to go away. to a great degree, it is an anachronism. what we need from the ils is a solid, business-process back end that would facilitate the functions of the library that are truly unique such as circulation, acquiring materials, and “cataloging” at the item level for what amounts to inventory-control purposes. most of the other tradi- tional ils functions could be rolled over into a central- ized system, like oclc, that would be cooperatively shared. the catalog itself should be treated as just another database in the world of resources we have access to. a single interface to those resources that would combine our local print holdings, electronic text (both journal and ebook), as well as multimedia material is what we should be demanding from our vendors. one book that needs to be required reading for all librarians, especially catalogers, is weinberger ’s everything is miscellaneous. he describes the three orders of order (self organization, metadata, and digi- tal); provides an extensive history of how western civilization has ordered information, specifically the links to nineteenth-century victorianism; and the con- cepts of lumping and splitting. in the end, weinberger argues that the digital environment allows users to manipulate information into their own organization sys- tem, disregarding all previous organizational attempts by supposed experts using outdated and outmoded systems. in the digital disorder of information, an object (leaf) can now be placed on many shelves (branches), figuratively speaking, and this new shape of knowledge brings out four strategic principles: . filter on the way out, not on the way in. . put each leaf on as many branches as possible. . everything is metadata and everything can be a label. . give up control. it is this last principle that libraries have challenges with. whether we agree with this principle or not, it has already happened. arguing about it, ignoring it, or just continuing to do business as usual isn’t going to change the fact that information is user-controled and user- initiated in the digital environment. so, where do we go from here? the future of technical services (and its staff) far be it from me to try to predict the future of libraries as viable, and more importantly marketable, information organizations in this new environment. one has only to examine the quotations from the first issues of technical services quarterly to see what happens to predictions and opinions. titles of some of the contributions (from , mind you) are worthy of mention: “library automation in the year ,” “musings on the future of the catalog,” and “libraries on the line.” there are developments, however, that require reexamination and strategic brain- storming regarding the future of library bibliographic organization and description. the appearance of worldcat local will have a tre- mendous impact on the disappearance of proprietary vendor opacs. there will no longer be a need for an integrated library system (ils); with worldcat local, the majority of the world’s marc bibliographic records are available in a library . format. the only things miss- ing are some type of inventory and acquisitions module that can be formatted locally and a circulation module. if oclc could focus their programming efforts on these two services and integrate them into worldcat local, library administrators and systems staff would no longer have to deal with proprietary and clunky opacs (and their huge budgetary lines), but could use the power of web . (and hopefully . ) tools and services to better position themselves in the new information marketplace. another major development is the google digitiza- tion project (and other associated ventures). while there are some concerns about quality and copyright, as well as issues related to the disappearance of print and the time involved to digitize all print, no one can deny the gradual and inevitable effect that mass digitization of print resources will have in the new information marketplace. just the fact that my research explorations for this article brought up digitized portions of the technical services quarterly articles is an example. more and more, published print information will be available in full-text online. what effect will this have on the physical collection that all libraries maintain, not only in terms of circulation, but also in terms of use of space, preservation, and collection devel- opment? no one knows for sure, but if the search strategies and information discovery patterns of our users are any information technology and libraries | june indication, then we need to be strategically preparing and developing directions and options. automatic metadata generation has been a topic of discussion for a number of years, and jane greenberg’s work at the university of north carolina–chapel hill is one of the leading examples of research in this area. while there are still viable concerns about metadata generation without any type of human intervention, semiautomatic and even nonlibrary-facilitated metadata generation has been successful in a number of venues. as libraries grapple with decreased budgets, multi- plying formats, fewer staff to do the work, and more retraining and reprofessional development of existing staff, library administrators have to examine all options to maximize personnel as well as budgetary resources. incorporating new technologies and tools for generat- ing metadata without human intervention into library workflows should be viewed as a viable option. user tagging would be included in this area. even intner, a long-time proponent of traditional technical services, has written that generating cataloging data automati- cally would be of great benefit to the profession, and that more tools and more programming ought to be focused toward this goal. so, with print workflows being replaced by digital and electronic workflows, how can administrators assist their technical services staff to remain viable in this new information environment? how can technical services staff not only help themselves but their supervisors and administrators to incorporate their unique talents, exper- tise, education, and experience toward the type of future scenarios indicated above? competencies and challenges for technical services staff there are some good opinions available for assisting technical services staff with moving into the new environ- ment. names have power, whether we like to admit it or not, and changing the name from “technical services” to something more understandable to our users, let alone our colleagues within the library, is one way to start. names such as “collections and data management services” or “reference data services” have been men- tioned. an interesting quote sums up the dilemma: it’s pretty clear that technical services departments have long been the ugly ducklings in the library pond, trumped by a quintet of swans: reference departments (the ones with answers for a grateful public); it depart- ments (the magicians who keep the computers hum- ming); children’s and youth departments (the warm and fuzzy nurturers); other specialty departments (the experts in good reads, music, art, law, business, medicine, government documents, av, rare books and manuscripts, you-name-it); and administrative groups (the big bosses). part of the trouble is that the rest of our colleagues don’t really know what technical services librarians do. they only know that we do it behind closed doors and talk about it in language no one else understands. if it can’t be seen, can’t be understood, and can’t be discussed, maybe it’s all smoke and mirrors, lacking real substance. it’s easy to ignore. ruschoff mentions competencies for technical ser- vices librarians in the new information environment: comfortable working in both print and digital worlds, specialized skills such as foreign languages and subject area expertise, comfortable working in both digital and web-based technologies (suggesting more computing and technology skills), expertise in digital asset manage- ment, and problem-solving analytical skills. in a recent blog posting summarizing a presentation at the ala annual conference on this topic, comparisons between catalogers going extinct or retooling are provided. the following is a summary of that post: converging trends ■■ more catalogers work at the support-staff level than as professional librarians. ■■ more cataloging records are selected by machines. ■■ more catalog records are being captured from pub- lisher data or other sources. ■■ more updating of catalog records is done via batch processes. ■■ libraries continue to deemphasize processing of sec- ondary research products in favor of unique primary materials. what are our choices? ■■ behind door number one—the extinction model. ■■ behind door number two—the retooling model. how it’s done ■■ extinction ■❏ keep cranking about how nobody appreciates us. ■❏ assert over and over that we’re already doing everything right—why should we change? ■❏ adopt a “chicken little” approach to envision- ing the future. ■■ retooling ■❏ considers what catalogers already do. ■❏ look for support. ■❏ find a new job. what catalogers do ■■ operate within the boundaries of detailed standards. ■■ describe items one-at-a-time. ■■ treat items as if they are intended to fit carefully from our readers | eden within a specific application—the catalog. ■■ ignore the rest of the world of information. what metadata librarians do ■■ think about descriptive data without preconceptions around descriptive level, granularity, or descriptive vocabularies. ■■ consider the entirety of the discovery and access issues around a set or collection of materials. ■■ consider users and uses beyond an individual ser- vice when making design decisions—not necessarily predetermined. ■■ leap tall buildings in a single bound. what new metadata librarians do ■■ be aware of changing user needs. ■■ understand the evolving information environment. ■■ work collaboratively with technical staff. ■■ be familiar with all metadata formats and encoding metadata. ■■ seek out tall buildings—otherwise jumping skills will atrophy. the cataloger skill set ■■ aacr , lc, etc. the metadata librarian skill set ■■ views data as collections, sets, streams. ■■ approaches the task as designing data to “play well with others.” characteristics of our new world ■■ no more ils ■■ bibliographic utilities are unlikely to be the central node for all data. ■■ creation of metadata will become more decentralized. ■■ nobody knows how this will all shake out, but meta- data librarians will be critical in forging solutions. while the above summary focuses on catalogers and their future, many of the directions also apply to any librarian or support staff member currently working in technical services. in a recent educause review article, brantley lists a number of mantras that all libraries need to repeat and keep in mind in this new information environment: ■■ libraries must be available everywhere. ■■ libraries must be designed to get better through use. ■■ libraries must be portable. ■■ libraries must know where they are. ■■ libraries must tell stories. ■■ libraries must help people learn. ■■ libraries must be tools of change. ■■ libraries must offer paths for exploration. ■■ libraries must help forge memory. ■■ libraries must speak for people. ■■ libraries must study the art of war. you will have to read the article to find out about that last point. the above mantras illustrate that each of these issues must also be aligned with the work done by technical services departments in support of the rest of the library’s services. and there definitely isn’t one right way to move forward; each library with its unique blend of services and staff has to define, initiate, and engender dialogue on change and strategic direction, and then actively make decisions with integrity and vigor toward both its users and its staff. as calhoun indicates, there are a number of challenges to feasibility for next steps in this area, some technically oriented but many based on our own organizational structures and strictures: ■■ difficulty achieving consensus on standardized, sim- plified, more automated workflows. ■■ unwillingness or inability to dispense with highly customized acquisitions and cataloging operations. ■■ overcoming the “not invented here” mindset pre- venting ready acceptance of cataloging copy from other libraries or external sources. ■■ resistance to simplifying cataloging. ■■ inability to find and successfully collaborate with necessary partners (e.g., ils vendors). ■■ difficulty achieving basic levels of system interoper- ability. ■■ slow development and implementation of necessary standards. ■■ library-centric decision making; inability to base priorities on how users behave and what they want ■■ limited availability of data to support management decisions. ■■ inadequate skill set among library staff; unwilling- ness or inability to retrain. ■■ resistance to change from faculty members, deans, or administrators. moving forward in the new information world in a recent discussion on the autocat electronic discus- sion list regarding the client-business paradigm now being impressed on library staff, an especially interesting quote puts the entire debate into perspective: the irony of this discussion is that our patrons/users/ clients [et al.] expect to be treated as well as business customers. they pay tuition or taxes to most of our institutions and expect to have a return in value. and a very large percentage of them care about the differ- ences between the government services vs. business information technology and libraries | june arguments we present. what they know is that when they want something, they want it. more library powers-that-be now come from the world of business rather than libraries because of the pressure on the bottom line. business administrators are viewed, even by those in public administration, as being more fiscally able than librarians. i would rec- ommend that we fuss less about titles and semantics and develop ways to show the value of libraries to the public. wheeler, in a recent educause review article, docu- ments a number of “eras” that colleges and universities have gone through in recent history. first is the “era of publishing,” followed by the “era of participation” with the appearance of the internet and its social networking tools. the next era, the “era of certitude,” is one in which users will want quick, timely answers to questions, along with some thought about the need and context of the question. wheeler espouses five dimensions that tools of certitude must have: reach, response, results, resources, and rights. he explains these dimensions in regards to var- ious tools and services that libraries can provide through human–human, human–machine, and machine–machine interaction. wheeler sees extensive rethinking and reengineering by libraries, campuses, and information technology to assist users to meet their information needs. are there ways that technical services staff can assist in these efforts? although somewhat dated, calhoun’s extensive article on what is needed from catalogers and librarians in the twenty-first century expounds a number of salient points. in table , she illustrates some of the many challenges fac- ing traditional library cataloging, providing her opinion on what the challenges are, why they exist, and some solutions for survivability and adaptability in the new marketplace. one quote in particular deserves attention: at the very least, adapting successfully to current demands will require new competencies for librarians, and i have made the case elsewhere that librarians must move beyond basic computer literacy to “it flu- ency”—that is, an understanding of the concepts of information technology, especially applying problem solving and critical thinking skills to using informa- tion technology. raising the bar of it fluency will be even more critical for metadata specialists, as they shift away from a focus on metadata production to approaches based on it tools and techniques on the one hand, and on consulting and teamwork on the other. as a result of the increasing need for it fluency among metadata specialists, they may become more closely allied with technical support groups in campus computing centers. the chief challenges for metadata spe- cialists will be getting out of library back rooms, becoming familiar with the larger world of university knowledge communities, and developing primary contacts with the appropriate domain experts and it specialists. getting out of the back room and interacting with users seems to be one of the dominant themes of evolv- ing technical services positions to fit the new information marketplace. putting web . tools and services into the library opac has also gained some momentum since the launch of the endeca-based opac at north carolina state university. as some people have stated, however, putting “lipstick on a pig” doesn’t change the fundamen- tal problems and poor usability of something that never worked well in the first place. in their recent article, jia mi and cathy weng tried to answer the following questions: why is the current opac ineffective? what can libraries and librarians do to deliver an opac that is as good as search engines to better serve our users? of course, the authors are biased toward the opac and wish to make it better, given that the last sentence in their abstract is, “revitalizing the opac is one of the press- ing issues that has to be accomplished.” users’ search patterns have already moved away from the opac as a discovery tool; why should personnel and resource investment continue to be allocated toward something that users have turned away from? in their recommenda- tions, mi and weng indicate that system limitations, not fully exploiting the functionality already made available by ilss, and the unsuitability of marc standards to online bibliographic display are the primary factors to the ineffectiveness of library opacs. exactly. debate and discussion on autocat after the publication of their article again shows the line drawn between conservative opin- ions (added value, noncommercialization, and overall ideals of the library profession and professional cata- loging workflows) and the newer push for open-source models, junking the opac, and learning and working with non-marc metadata standards and tools. conclusion from an administrative point of view, there are a number of viable options for making technical services as efficient as possible, in its current emanation: ■■ conduct a process review of all current workflows, following each type of format from receipt at loading dock to access by user. revise and redesign work- flows for efficiency. ■■ eliminate all backlogs, incorporating and standardiz- ing various types of bibliographic organization (from brief records to full records, using established criteria of importance and access). ■■ as much as possible, contract with vendors to make from our readers | eden all print materials shelf-ready, establishing and moni- toring profiles for quality and accuracy. establish a rate of error that is amenable to technical services staff; once that error rate is met, review incoming print materials only once or twice a year. ■■ assure technical services staff that their skills, expe- rience, and attention to detail are needed in the electronic environment, and provide training and professional development to assist them in scan- ning and digitizing unique collections, learning non-marc metadata standards, improving project management, and performing consultation training to interact with faculty and students who work with data sets, metadata, and research planning. support and actively work for revised job reclassification of library support staff positions. most libraries are forced to work with fewer staff, and it is essential that current personnel are valued for their institutional knowledge and skill sets (knowledge man- agement philosophy). library administrations need to emphasize to their staff that the organization has a vested interest in providing them with the tools and training they need to assist the organization in the new informa- tion marketplace. the status quo of technical services operations is no longer viable or cost-effective; all of us must look at ways to regain market share and restruc- ture our organizations to collaborate and consult with users regarding their information and research needs. no longer is it enough to just provide access to information; we must also provide tools and assistance to the user in manipulating that information. to end, i would like to quote from a few of the articles from that issue of technical services quarterly i have alluded to throughout this chapter: like all prognostications, predictions about cataloging in a fully automated library may bear little resem- blance to the ultimate reality. while the future cata- loging scenario discussed here may seem reasonable now, it could prove embarrassing to read – years hence. still, i would be pleasantly surprised if, by the year , ts operations are not fully integrated, ts staff has not been greatly reduced, there has not been a large-scale jump in ts productivity accompanied by a dramatic decline in ts costs, and if most of us are not cooperating through a national database. in conclusion, i will revert to my first subject, the uncertain nature of predictions. in addition to the fear- less predictions already recorded, i predict that some of these predictions will come true and perhaps even most of them. some of them will come true, but not in the time anticipated, while others never will. let us hope that the influences not guessed that will prevent the actualization of some of these predictions will be happy ones, not dire. however they turn out, i predict that in ten years no one will remember or really care what these predictions were. technical services as we know them now may well not exist by the end of the century. the aims of technical services will exist for as long as there are libraries. the technical services quarterly may well have changed its name and its coverage long before then, but its con- cerns will remain real and the work to which many of us devote our lives will remain worthwhile. there can be few things in life that are as worth doing as enabling libraries to fulfill their unique and uniquely important role in culture and civilization. twenty-five years have come and gone; some of the predictions in this first issue of technical services quarterly came true, many of them did not. there have been dra- matic changes in those twenty-five years, most of which were unforeseen, as they always are. what is a certainty is that libraries can no longer sustain or maintain the status quo in technical services. what also is a certainty is that technical services staff, with their unique skills, talents, abilities, and knowledge in relation to the organization and description of information, are desperately needed in the new information environment. it is the responsibil- ity of both library administrators and technical services staff to work together to evolve and redesign workflows, standards, procedures, and even themselves to survive and succeed into the future. references . norman d. stevens, “selections from a dictionary of libinfosci terms,” in “beyond ‘ ’: the future of technical services,” special issue, technical services quarterly , no. – (fall/winter ): . . marcia j. bates, “improving user access to library catalog and portal information: final report,” (paper pre- sented at the library of congress bicentennial conference on bibliographic control for the new millennium, june , ): , http://www.loc.gov/catdir/bibcontrol/ . batesreport - .doc.pdf (accessed apr. , ). see also karen calhoun, “the changing nature of the catalog and its integration with other discovery tools,” final report to the library of congress, mar. , , , http://www.loc.gov/catdir/calhoun-report-final .pdf (accessed apr. , ). . university of california libraries bibliographic services task force, “rethinking how we provide bibliographic ser- vices for the university of california,” final report, dec. , , http://libraries.universityofcalifornia.edu/sopag/bstf/final. pdf (accessed apr. , ). . deanna b. marcum, “the future of cataloging,” library resources & technical services , no. (jan. ): , http://www .loc.gov/library/reports/catalogingspeech.pdf (accessed apr. information technology and libraries | june , ). . karen calhoun, “being a librarian: metadata and meta- data specialists in the twenty-first century,” library hi tech , no. ( ), http://www.emeraldinsight.com/insight/view contentservlet?filename=published/emeraldfulltextarticle/ articles/ .html (accessed apr. , ). . calhoun, “the changing nature of the catalog,” . . bradford lee eden, “ending the status quo,” american libraries , no. (mar. ): ; eden, introduction to “infor- mation organization future for libraries,” library technology reports , no. (nov./dec. ): – . . see karen schneider’s “how opacs suck” series on the ala techsource blog, http://www.techsource.ala.org/ blog/ / /how-opacs-suck-part- -relevance-rank-or-the -lack-of-it.html, http://www.techsource.ala.org/blog/ / / how-opacs-suck-part- -the-checklist-of-shame.html, and http:// www.techsource.ala.org/blog/ / /how-opacs-suck-part- -the-big-picture.html (accessed apr. , ). . h. frank cervone, quoted in ellen bahr, “dreaming of a better ils,” computers in libraries , no. (oct. ): . . david weinberger, everything is miscellaneous: the power of the new digital disorder (new york: times, ). . for a list of these concerns, see robert darnton, “the library in the new age,” the new york review of books , no. (june , ), http://www.nybooks.com/articles/ (accessed apr. , ). . see calhoun, “the changing nature of the catalog,” . . see the metadata research center, “automatic metadata generation applications (amega),” http://ils.unc.edu/mrc/ amega (accessed, apr. , ). . sheila s. intner, “generating cataloging data automati- cally,” technicalities , no. (mar./apr. ): , – . . sheila s. intner, “a technical services makeover,” techni- calities , no. (sept./oct. ): , – . . ibid, (emphasis added). . carlen ruschoff, “competencies for st century techni- cal services,” technicalities , no. (nov./dec. ): , – . . diane hillman, “a has-been cataloger looks at what cataloging will be,” online posting, metadata blog, july , , http://blogs.ala.org/nrmig.php?title=creating_the_future_of_ the_catalog_aamp_&more= &c= &tb= &pb= (accessed apr. , ). . peter brantley, “architectures for collaboration: roles and expectations for digital libraries,” educause review , no. (mar./apr. ): – . . calhoun, “the changing nature of the catalog,” . . brian briscoe, “that business/customer stuff (was: let- ter to al),” online posting, autocat, may , . . brad wheeler, “in search of certitude,” educause review , no. (may/june ): – . . ibid., . . karen calhoun, “being a librarian.” . ibid. . ibid. (emphasis added). . andrew pace, quoted in roy tennant, “digitl librar- ies: ‘lipstick on a pig,’” library journal, apr. , , http:// www.libraryjournal.com/article/ca .html (accessed apr. , ). . jia mi and cathy weng, “revitalizing the library opac: interface, searching, and display challenges,” information tech- nology & libraries , no. (mar. ): – . . gregor a. preston, “how will automation affect cata- loging staff?” in “beyond ‘ ’: the future of technical ser- vices,” special issue, technical services quarterly , no. – (fall/ winter ): . . david c. taylor, “the library future: computers,” in “beyond ‘ ’: the future of technical services,” special issue, technical services quarterly , no. – (fall/winter ): – . . michael gorman, “technical services, – (and before),” in “beyond ‘ ’: the future of technical services,” special issue, technical services quarterly , no. – (fall/winter ): . lita cover , cover neal-schuman cover index to advertisers information technology and libraries | march the path toward global interoperability in cataloging ilana tolkoff libraries began in complete isolation with no uniformity of standards and have grown over time to be ever more interoperable. this paper examines the current steps toward the goal of universal interoperability. these projects aim to reconcile linguistic and organizational obstacles, with a particular focus on subject headings, name authorities, and titles. i n classical and medieval times, library catalogs were completely isolated from each other and idiosyncratic. since then, there has been a trend to move toward greater interoperability. we have not yet attained this international standardization in cataloging, and there are currently many challenges that stand in the way of this goal. this paper will examine the teleological evolution of cataloging and analyze the obstacles that stand in the way of complete interoperability, how they may be overcome, and which may remain. this paper will not provide a comprehensive list of all issues pertaining to interoper- ability; rather, it will attempt to shed light on those issues most salient to the discussion. unlike the libraries we are familiar with today, medi- eval libraries worked in near total isolation. most were maintained by monks in monasteries, and any regulations in cataloging practice were established by each religious order. one reason for their lack of regulations was that their collections were small by our standards; a monastic library had at most a few hundred volumes (a couple thousand in some very rare cases). the “armarius,” or librarian, kept more of an inventory than an actual cata- log, along with the inventories of all other valuable pos- sessions of the monastery. there were no standard rules for this inventory-keeping, although the armarius usually wrote down the author and title, or incipit if there was no author or title. some of these inventories also contained bibliographic descriptions, which most often described the physical book rather than its contents. the inventories were usually taken according to the shelf organization, which was occasionally based on subject, like most librar- ies are today. these trends in medieval cataloging varied widely from library to library, and their inventories were entirely different from our modern opacs. the inventory did not provide users access to the materials. instead, the user consulted the armarius, who usually knew the col- lection by heart. this was a reasonable request given the small size of the collections. this type of nonstandardized cataloging remained relatively unchanged until the nineteenth century, when charles c. jewett introduced the idea of a union catalog. jewett also proposed having stereotype plates for each bibliographic record, rather than a book catalog, because this could reduce costs, create uniformity, and organize records alphabetically. this was the precursor to the twentieth-century card catalog. while many of jewett’s ideas were not actually practiced during his lifetime, they laid the foundation for later cataloging practices. the twentieth century brought a great revolution in cataloging standards, particularly in the united states. in , the library of congress subject headings (lcsh) were first published and introduced a controlled vocabu- lary to american cataloging. the s saw a wide array of advancements in standardization. the library of congress (lc) developed marc, which became a national standard in . it also was the time of the cre- ation of anglo-american cataloguing rules (aacr), the paris principles, and international standard bibliographic description (isbd). while many of these standardization projects were uniquely american or british phenomena, they quickly spread to other parts of the world, often in translated versions. while the technology did not yet exist in the s to provide widespread local online catalogs, technology did allow for union catalogs containing the records of many libraries in a single database. these union catalogs included the research libraries information network (rlin), the oclc online computer library center (oclc), and the western library network (wln). in the s the local online public access catalog (opac) emerged, and in the s opacs migrated to the web (webpacs). currently, most libraries have opacs and are members of oclc, the largest union catalog, used by more than , libraries in countries and ter- ritories. now that most of the world’s libraries are on oclc, librarians face the challenge and inconvenience of dis- crepancies in cataloging practice due to the differing stan- dards of diverse countries, languages, and alphabets. the fields of language engineering and linguistics are work- ing on various language translation and analysis tools. some of these include machine translation; ontology, or the hierarchical organization of concepts; information extraction, which deciphers conceptual information from unorganized information, such as that on the web; text summarization, in which computers create a short sum- mary from a long piece of text; and speech processing, which is the computer analysis of human speech. while these are all exciting advances in information technol- ogy, as of yet they are not intelligent enough to help us establish cataloging interoperability. it will be interesting to see whether language engineering tools will be capable of helping catalogers in the future, but for now they are ilana tolkoff (ilana.tolkoff@gmail.com) holds a ba in music and italian from vassar college, an ma in musicology from brandeis university, and an mls from the university at buffalo. she is cur- rently seeking employment as a music librarian. the path toward global interoperability in cataloging | tolkoff best at making sense of unstructured information, such as the web. the interoperability of library catalogs, which consist of highly structured information, must be tackled through software that innovative librarians of the future will produce. in an ideal world, oclc would be smoothly interop- erable at a global level. a single thesaurus of subject headings would have translations in every language. there would be just one set of authority files. all mani- festations of a single work would be grouped under the same title, translatable to all languages. there would be a single bibliographic record for a single work, rather than multiple bibliographic records in different languages for the same work. this single bibliographic record could be translatable into any language, so that when searching in worldcat, one could change the settings to any language to retrieve records that would display in that chosen lan- guage. when catalogers contribute to oclc, they would create the records in their respective languages, and once in the database the records would be translatable to any other language. because records would be so fluidly translatable, an opac could be searched in any language. for example, the default settings for the university at buffalo’s opac could be english, but patrons could change those settings to accommodate the great variety of international students doing research. this vision is uto- pian to say the least, and it is doubtful that we will ever reach this point. but it is valuable to establish an ideal scenario to aim our innovation in the right direction. one major obstacle in the way of global interoper- ability is the existence of different alphabets and the inherently imperfect nature of transliteration. there are essentially two types of transliteration schemes: those based on phonetic structure and those based on mor- phemic structure. the danger of phonetic transliteration, which mimics pronunciation, is that semantics often get lost. it fails to differentiate between homographs (words that are spelled and pronounced the same way but have different meanings). complications also arise when there are differences between careful and casual styles of speech. park asserts, “when catalogers transcribe words according to pronunciation, they can create inconsistent and arbitrary records.” morphemic transliteration, on the other hand, is based on the meanings of morphemes, and sometimes ends up being very different from the pronunciation in the source language. one advantage to this, however, is that it requires fewer diacritics than phonetic transliteration. park, whose primary focus is on korean–roman transliteration, argues that the mccune reischauer phonetic transliteration that libraries use loses too much of the original meaning. in other alphabets, however, phonetic transliteration may be more beneficial, as in the lc’s recent switch to pinyin transliteration in chinese. the lc found pinyin to be more easily search- able than wade-giles or monosyllabic pinyin, which are both morphemic. however, another problem with translit- eration that neither phonetic nor morphemic schemes can solve is word segmentation—how a transliterated word is divided. this becomes problematic when there are no contextual clues, such as in a bibliographic record. other obstacles that stand in the way of interoperabil- ity are the diverse systems of subject headings, author- ity headings, and titles found internationally. resource description and access (rda) will not deal with subject headings because it is such a hefty task, so it is unlikely that subject headings will become globally interoperable in the near future. fortunately, twenty-four national libraries of english speaking countries use lcsh, and twelve non-english-speaking countries use a translated or modified version of lcsh. this still leaves many more countries that use their own systems of subject headings, which ultimately need to be made interoperable. even within a single language, subject headings can be compli- cated and inconsistent because they can be expressed as a single noun, compound noun, noun phrase, or inverted phrase; the problem becomes even greater when trying to translate these to other languages. bennett, lavoie, and o’neill note that catalogers often assign different subject headings (and classifications) to different manifestations of the same work. that is, the record for the novel gone with the wind might have different subject headings than the record for the movie. this problem could poten- tially be resolved by the functional requirements for bibliographic records (frbr), which will be discussed below. translation is a difficult task, particularly in the con- text of strict cataloging rules. it is especially complicated to translate among unrelated languages, where one might be syntactic and the other inflectional. this means that there are discrepancies in the use of prepositions, con- junctions, articles, and inflections. the ability to add or remove terms in translation creates endless variations. a single concept can be expressed in a morpheme, a word, a phrase, or a clause, depending on the language. there also are cultural differences that are reflected in differ- ent languages. park gives the example of how anglo- american culture often names buildings and brand names after people, reflecting our culture’s values of individualism, while in korea this phenomenon does not exist at all. on the other hand, korean’s use of formal and informal inflections reflects their collectivist hierarchical culture. another concept that does not cross cultural lines is the korean pumasi system in which family and friends help someone in a time of need with the understanding that the favor will be returned when they need it. this cannot be translated into a single english word, phrase, or subject heading. one way of resolving ambiguity in translations is through modifiers or scope notes, but this is only a partial solution. because translation and transliteration are so difficult, information technology and libraries | march as well as labor-intensive, the current trend is to link already existing systems. multilingual access to subjects (macs) is one such linking project that aims to link subject headings in english, french, and german. it is a joint project under the conference of european national librarians among the swiss national library, the bibliothèque nationale de france (bnf), the british library (bl), and die deutsche bibliothek (ddb). it aims to link the english lcsh, the french répertoire d’autorité matière encyclopédique et alphabétique unifié (rameau), and the german schlagwortnormdatei/ regeln für den schlagwortkatalog (swd/rswk). this requires manually analyzing and matching the concepts in each heading. if there is no conceptual equivalent, then it simply stands alone. macs can link between headings and strings or even create new headings for linking pur- poses. this is not as fruitful as it sounds, however, as there are fewer correspondences than one might expect. the macs team experimented with finding correspondences by choosing two topics: sports, which was expected to have a particularly high number of correspondences, and theater, which was expected to have a particularly low number of correspondences. of the sports head- ings, percent matched in all three languages, percent matched in two, and percent was unmatched. of the theater headings, percent matched in three lan- guages, percent matched in two, and percent was unmatched. even in the most cross-cultural subject of sports, percent of terms did not correspond fully, mak- ing one wonder whether linking will work well enough to prevail. a similar project—the virtual international authority file (viaf)—is being undertaken for authority headings, a joint project of the lc, the bnf, and ddb, and now including several other national libraries. viaf aims to link (not consolidate) existing authority files, and its beta version (available at http://viaf.org) allows one to search by name, preferred name, or title. oclc’s software mines these authority files and the titles associated with them for language, lc control number, lc classifica- tion, usage, title, publisher, place of publication, date of publication, material type, and authors. it then derives a new enhanced authority record, which facilitates map- ping among authority records in all of viaf’s languages. these derived authority records are stored on oai serv- ers, where they are maintained and can be accessed by users. users can search viaf by a single national library or broaden their possibilities by searching all participat- ing national libraries. as of , between the lc’s and ddb’s authority files, there were , matches, includ- ing , complex matches (one-to-many), and , unique matches (one-to-one) out of , , lc names and , , ddb names. ultimately, viaf could be used for still more languages, including non-roman alphabets. recently the national library of israel has joined, and viaf can link to the hebrew alphabet. a similar project to viaf that also aimed to link authority files was linking and exploring authority files (leaf), which was under the auspices of the information society technologies programme of the fifth framework of the european commission. the three-year project began in with dozens of libraries and organizations (many of which are national libraries), representing eight languages. its website describes the project as follows: information which is retrieved as a result of a query will be stored in a pan-european “central name authority file.” this file will grow with each query and at the same time will reflect what data records are rel- evant to the leaf users. libraries and archives want- ing to improve authority information will thus be able to prioritise their editing work. registered users will be able to post annotations to particular data records in the leaf system, to search for annotations, and to download records in various formats. park identifies two main problems with linking authority files. one is that name authorities still contain some language-specific features. the other is that disam- biguation can vary among name authority systems (e.g., birth/death dates, corporate qualifiers, and profession/ activity). these are the challenges that projects like leaf and viaf must overcome. while the linking of subject headings and name authorities is still experimental and imperfect, the frbr model for linking titles is much more promising and will be incorporated in the soon-to-be-released rda. according to bennett, lavoie, and o’neill, there are three important benefits to frbr: ( ) it allows for different views of a bibliographic database, ( ) it creates a hierarchy of bibliographic entities in the catalog such that all versions of the same work fall into a single collapsible entry point, ( ) and the confluence of the first two benefits makes the cata- log more efficient. in the frbr model, the bibliographic record consists of four entities: ( ) the work, ( ) the expres- sion, ( ) the manifestation, and ( ) the item. all manifesta- tions of a single work are grouped together, allowing for a more economical use of information because the title needs to be entered only once. that is, a “title authority file” will exist much like a name authority file. this means that all editions in all languages and in all formats would be grouped under the same title. for example, the lord of the rings title would include all novels, films, translations, and editions in one grouping. this would reduce the number of bibliographic records, and as danskin notes, “the idea of creating more records at a time when publishing output threatens to outstrip the cataloguing capacity of national bibliographic agencies is alarming.” the frbr model is particularly beneficial for com- plex canonical works like the bible. there are a small number of complex canonical works, but they take up a the path toward global interoperability in cataloging | tolkoff disproportionate number of holdings in oclc. because this only applies to a small number of works, it would not be difficult to implement, and there would be a disproportionate benefit in the long run. there is some uncertainty, however, in what constitutes a complex work and whether certain items should be grouped under the same title. for instance, should prokofiev’s romeo and juliet be grouped with shakespeare’s? the advantage of the frbr model for titles over subject headings or name authorities is that no such thing as a title authority file exists (as conceptualized by frbr). we would be able to start from scratch, creating such title authority files at the international level. subject headings and name authori- ties, on the other hand, already exist in many different forms and languages so that cross-linking projects like viaf might be our only option. it is encouraging to see the strides being made to make subject headings, name authority headings, and titles globally interoperable, but what about other access points within a record’s bibliographic description? these are usually in only one language, or two if cataloged in a bilingual country. should these elements (format, contents, and so on) be cross-linked as well, and is this even possible? what should reasonably be considered an access point? most people search by subject, author, or title, so perhaps it is not worth making other types of access points interoperable for the few occasions when they are useful. yet if percent universal interoperabil- ity is our ultimate utopian goal, perhaps we should not settle for anything less than true international access to all fields in a record. because translation and transliteration are such com- plex undertakings, linking of extant files is the future of the field. there are advantages and disadvantages to this. on the one hand, linking these files is certainly bet- ter than having them exist only for their own countries. they are easily executed projects that would not require a total overhaul of the way things currently stand. the disadvantages are not to be ignored, however. the fact that files do not correspond perfectly from language to language means that many files will remain in isolation in the national library that created them. another problem is that cross-linking is potentially more confusing to the user; the search results on http://www.viaf.org are not always simple and straightforward. if cross-linking is where we are headed, then we need to focus on a more user-friendly interface. if the ultimate goal of interoper- ability is simplification, then we need to actually simplify the way query results are organized rather than make them more confusing. very soon rda will be released and will bring us to a new level of interoperability. aacr arrived in , and though it has been revised several times, it is in many ways outdated and mainly applies to books. rda will bring something completely new to the table. it will be flexible enough to be used in other metadata schemes besides marc, and it can even be used by different industries such as publishers, museums, and archives. its incorporation of the frbr model is exciting as well. still, there are some practical problems in implementing rda and frbr, one of which is that reeducating librar- ians about the new rules will be costly and take time. also, frbr in its ideal form would require a major over- haul of the way oclc and integrated library systems currently operate, so it will be interesting to see to what extent rda will actually incorporate frbr and how it will be practically implemented. danskin asks, “will the benefits of international co-operation outweigh the costs of effecting changes? is the usa prepared to change its own practices, if necessary, to conform to european or wider ifla standards?” it seems that the united states is in fact ready and willing to adopt frbr, but to what extent is yet to be determined. what i have discussed in this paper are some of the more prominent international standardization projects, although there are countless others, such as eurowordnet, the open language archives community (olac), and international cataloguing code (icc), to name but a few. in general, the current major projects consist of linking subject headings, name authority files, and titles in multiple languages. linking may not have the best cor- respondence rates, we have still not begun to tackle the cross-linking of other bibliographic elements, and at this point search results may be more confusing than help- ful. but the existence of these linking projects means we are at least headed in the right direction. the emergent universality of oclc was our most recent step toward interoperability, and it looks as if cross-linking is our next step. only time will tell what steps will follow. references . lawrence s. guthrie ii, “an overview of medieval library cataloging,” cataloging & classification quarterly , no. ( ): – . . lois mai chan and theodora hodges, cataloging and classification: an introduction, rd ed. (lanham, md.: scarecrow, ): . . ibid., – . . ibid., – . . oclc, “about oclc,” http://www.oclc.org/us/en/ about/default.htm (accessed dec. , ). . jung-ran park, “cross-lingual name and subject access: mechanisms and challenges,” library resources & technical ser- vices , no. ( ): . . ibid., . . ibid. continued on page tagging: an organization scheme for the internet | visser international and o’reilly media, web . refers to the web as being a platform for harnessing the collective power of internet users interested in creating and sharing ideas and information without mediation from corporate, government, or other hierar- chical policy influencers or regulators. web . is a much more fluid concept as of this writing. there are individuals who use it to refer to a semantic web where information is analyzed or processed by software designed specifically for computers to carry out the currently human-mediated activity of assigning meaning to information on a webpage. there are librarians involved with exploring virtual-world librarianship who refer to the d environment as web . . the important point here is that what internet users now know as web . is in the process of being altered by individuals continually experimenting with and improving upon existing web applications. web . is the undefined future of the participatory internet. . clay shirky, “here comes everybody: the power of organizing without organizations” (presentation videocast, berkman center for internet & society, harvard university, cambridge, mass., ), http://cyber.law.harvard.edu/inter active/events/ / /shirky (accessed oct. , ). . ibid. . lawerence lessig, “early creative commons history, my version,” videocast, aug. , , lessig . , http://lessig.org/ blog/ / /early_creative_commons_history.html (accessed aug. , ). . elaine peterson, “beneath the metadata: some philosophi- cal problems with folksonomy,” d-lib magazine , no. ( ), http://www.dlib.org/dlib/november /peterson/ peterson .html (accessed sept. , ). . clay shirky, “ontology is overrated: categories, links, and tags” online posting, spring , clay shirky’s writings about the internet, http://www.shirky.com/writings/ontology_ overrated.html#mind_reading (accessed sept. , ). . gene smith, tagging: people-powered metadata for the social web (berkeley, calif.: new riders, ): . . ibid., . . thomas vander wal, “folksonomy,” online posting, feb. , , vanderwal.net, http://www.vanderwal.net/folksonomy .html (accessed aug. , ). . thomas vander wal, “explaining and showing broad and narrow folksonomies,” online posting, feb. , , personal infocloud, http://www.personalinfocloud.com/ / / explaining_and_.html (accessed aug. , ). . shirky, “ontology is overrated.” . ibid. . michael arrington, “exclusive: screen shots and feature overview of delicious . preview,” online posting, june , , techcrunch, http://www.techcrunch.com/ / / / exclusive-screen-shots-and-feature-overview-of-delicious- -preview/(accessed jan. , ). . smith, tagging, – . . vander wal, “explaining and showing broad and narrow folksonomies.” . adam mathes, “folksonomies—cooperative classifica- tion and communication through shared metadata” (graduate paper, university of illinois urbana–champaign, dec. ); peterson, “beneath the metadata”; shirky, “ontology is over- rated”; thomas and griffin, “who will create the metadata for the internet?” . shirky, “ontology is overrated.” . peterson, “beneath the metadata.” . cory doctorow, “metacrap: putting the torch to seven straw-men of the meta-utopia,” online posting, aug. , , the well, http://www.well.com/~doctorow/metacrap.htm (accessed sept. , ). . marieke guy and emma tonkin, “folksonomies: tidy- ing up tags?” d-lib magazine , no. ( ), http://www.dlib .org/dlib/january /guy/ guy.html (accessed sept. , ). . shirky, “ontology is overrated.” global interoperability continued from page . julie renee moore, “rda: new cataloging rules, com- ing soon to a library near you!” library hi tech news , no. , ( ): . . rick bennett, brian f. lavoie, and edward t. o’neill, “the concept of a work in worldcat: an application of frbr,” library collections, acquisitions, & technical services , no. , ( ): . . park, “cross-lingual name and subject access.” . ibid. . thomas b. hickey, “virtual international authority file” (microsoft powerpoint presentation, ala annual conference, new orleans, june ), http://www.oclc.org/research/ projects/viaf/ala c.ppt (accessed dec. , ). . leaf, “leaf project consortium,” http://www.crxnet .com/leaf/index.html (accessed dec. , ). . bennett, lavoie, and o’neill, “the concept of a work in worldcat.” . alan danskin, “mature consideration: developing biblio- graphic standards and maintaining values,” new library world , no. / , ( ): . . ibid. . bennett, lavoie, and o’neill, “the concept of a work in worldcat.” . moore, “rda.” . danskin, “mature consideration,” . . ibid.; park, “cross-lingual name and subject access.” reducing psychological resistance to digital repositories | quinn and mit mandates, and other mandates such as the one instituted at stanford’s school of education, have come to pass, and the registry of open access repository material archiving policies (roarmap) lists more than mandates around the world that now exist. while it is too early to tell whether these developments will be successful in getting faculty to deposit their work in digital repositories, they at least establish a precedent that other institutions may follow. how many institutions fol- low and how effective the mandates will be once enacted remains to be seen. will all colleges and universities, or even a majority, adopt mandates that require faculty to deposit their work in repositories? what of those that do not? even if most institutions are successful in instituting mandates, will they be sufficient to obtain faculty cooperation? for those institutions that do not adopt mandates, how are they going to persuade faculty to participate in self-archiving, or even in some variation—such as having surrogates (librarians, staff, or graduate assistants) archive the work of faculty? are mandates the only way to ensure faculty cooperation and compliance, or are mandates even neces- sarily the best way? to begin to adequately address the problem of user resistance to digital repositories, it might help to first gain some insight into the psychology of resistance. the existing literature on user behavior with regard to digital repositories devotes scant attention to the psy- chology of resistance. in an article entitled “institutional repositories: partnering with faculty to enhance scholarly communication,” johnson discusses the inertia of the traditional publishing paradigm. he notes that this inertia is most evident in academic faculty. this would suggest that the problem of eliciting user cooperation is primarily motivational and that the problem is more one of indifference than active resistance. heterick, in his article “faculty attitudes toward electronic resources,” suggests that one reason faculty may be resistant to digital repositories is because they do not fully trust them. in response to a survey he conducted, percent of faculty felt that libraries should maintain paper archives. the implication is that digital reposi- tories and archives may never completely replace hard copies in the minds of scholars. in “understanding faculty to improve content recruitment for institutional repositories,” foster and gibbons point out that faculty complain of having too much work already. they resent any additional work that contributing to a digital repository might entail. thus the authors echo johnson in suggesting that faculty resistance the potential value of digital repositories is dependent on the cooperation of scholars to deposit their work. although many researchers have been resistant to sub- mitting their work, the literature on digital repositories contains very little research on the psychology of resis- tance. this article looks at the psychological literature on resistance and explores what its implications might be for reducing the resistance of scholars to submitting their work to digital repositories. psychologists have devised many potentially useful strategies for reducing resistance that might be used to address the problem; this article examines these strategies and how they might be applied. o bserving the development and growth of digital repositories in recent years has been a bit like rid- ing an emotional roller coaster. even the definition of what constitutes a repository may not be the subject of complete agreement, but for the purposes of this study, a repository is defined as an online database of digital or digitized scholarly works constructed for the purpose of preserving and disseminating scholarly research. the initial enthusiasm expressed by librarians and advocates of open access toward the potential of repositories to make significant amounts of scholarly research avail- able to anyone with internet access gradually gave way to a more somber appraisal of the prospects of getting faculty and researchers to deposit their work. in august , bailey posted an entry to his digital koans blog titled “institutional repositories: doa?” in which he noted that building digital repository collections would be a long, arduous, and costly process. the success of repositories, in his view, will be a function not so much of technical considerations as of attitudinal ones. faculty remain unconvinced that repositories are important, and there is a critical need for outreach programs that point to repositories as an important step in solving the crisis in scholarly communication. salo elaborated on bailey’s post with “yes, irs are broken. let’s talk about it,” on her own blog, caveat lector. salo points out that institutional repositories have not fulfilled their early promise of attracting a large num- ber of faculty who are willing to submit their work. she criticizes repositories for monopolizing the time of library faculty and staff, and she states her belief that repositories will not work without deposit mandates, but that man- dates are impractical. subsequent events in the world of scholarly com- munication might suggest that mandates may be less impractical than salo originally thought. since her post, the national institutes of health mandate, the harvard brian quinn (brian.quinn@ttu.edu) is social sciences librarian, texas tech university libraries, lubbock. brian quinn reducing psychological resistance to digital repositories information technology and libraries | june whether or not this was actually the case. this study also suggests that a combination of both cognitive and affective processes feed faculty resistance to digital repositories. it can be seen from the preceding review of the lit- erature that several factors have been identified as being possible sources of user resistance to digital repositories. yet the authors offer little in the way of strategies for addressing this resistance other than to suggest work- around solutions such as having nonscholars (e.g., librarians, graduate students, or clerical staff) serve as proxy for faculty and deposit their work for them, or to suggest that institutions mandate that faculty deposit their work. similarly, although numerous arguments have been made in favor of digital repositories and open access, they do not directly address the resistance issue. in contrast, psychologists have studied user resistance extensively and accumulated a body of research that may suggest ways to reduce resistance rather than try to circumvent it. it may be helpful to examine some of these studies to see what insights they might offer to help address the problem of user resistance. it should be pointed out that resistance as a topic has been addressed in the business and organizational literature, but has generally been approached from the standpoint of management and organizational change. this study has chosen to focus primarily on the psychol- ogy of resistance because many repositories are situated in a university setting. unlike employees of a corporation, faculty members typically have a greater degree of auton- omy and latitude in deciding whether to accommodate new work processes and procedures into their existing routines, and the locus of change will therefore be more at an individual level. ■■ the psychology of user resistance psychologists define resistance as a preexisting state or attitude in which the user is motivated to counter any attempts at persuasion. this motivation may occur on a cognitive, affective, or behavioral level. psychologists thus distinguish between a state of not being persuaded and one in which there is actual motivation to not com- ply. the source of the motivation is usually an affective state, such as anxiety or ambivalence, which itself may result from cognitive problems, such as misunderstand- ing, ignorance, or confusion. it is interesting to note that psychologists have long viewed inertia as one form of resistance, suggesting paradoxically that a person can be motivated to inaction. resistance may also manifest itself in more subtle forms that shade into indifference, suspicion of new work processes or technologies, and contentment with the status quo. may be attributed at least in part to motivation. in another article published a few months later, foster and gibbons suggest that the main reason faculty have been slow to deposit their work in digital repositories is a cog- nitive one: faculty have not understood how they would benefit by doing so. the authors also mention that users may feel anxiety when executing the sequence of techni- cal steps needed to deposit their work, and that they may also worry about possible copyright infringement. the psychology of resistance may thus manifest itself in both cognitive and affective ways. harley and her colleagues talk about faculty not perceiving any reward for depositing their work in their article “the influence of academic values on scholarly publication and communication practices.” this percep- tion results in reduced drive to participate. anxiety is another factor contributing to resistance: faculty fear that their work may be vulnerable to plagiarism in an open- access environment. in “towards user responsive institutional repositories: a case study,” devakos suggests that one source of user resistance is cognitive in origin. scholars do not submit their work frequently enough to be able to navigate the interface from memory, so they must reinitiate the learning process each time they submit their work. the same is true for entering metadata for their work. their sense of control may also be threatened by any limitations that may be imposed on substituting later iterations of their work for earlier versions. davis and connolly point to several sources of con- fusion, uncertainty, and anxiety among faculty in their article “institutional repositories: evaluating the reasons for non-use of cornell university’s installation of dspace.” cognitive problems arise from having to learn new technology to deposit work and not knowing copy- right details well enough to know whether publishers would permit the deposit of research prior to publica- tion. faculty wonder whether this might jeopardize their chances of acceptance by important journals whose edi- tors might view deposit as a form of prior publication that would disqualify them from consideration. there is also fear that the complex structure of a large repository may actually make a scholar’s work more difficult to find; fac- ulty may not understand that repositories are not isolated institutional entities but are usually searchable by major search engines like google. kim also identifies anxiety about plagiarism and confusion about copyright as being sources of faculty resistance in the article “motivating and impeding factors affecting faculty contribution to institutional repositories.” kim found that plagiarism anxiety made some faculty only willing to deposit already-published work and that prepublication material was considered too risky. faculty with no self-archiving experience also felt that many publishers do not allow self-archiving, reducing psychological resistance to digital repositories | quinn more open to information that challenges their beliefs and attitudes and are more open to suggestion. thus before beginning a discussion of why users should deposit their research in repositories, it might help to first affirm the users’ self-concept. this could be done, for example, by reminding them of how unbiased they are in their work or how important it is in their work to be open to new ideas and new approaches, or how successful they have been in their work as scholars. the affirmation should be subtle and not directly related to the repository situation, but it should remind them that they are open- minded individuals who are not bound by tradition and that part of their success is attributable to their flexibility and adaptability. once the users have been affirmed, librar- ians can then lead into a discussion of the importance of submitting scholarly research to repositories. self-generated affirmations may be even more effec- tive. for example, another way to affirm the self would be to ask users to recall instances in which they successfully took a new approach or otherwise broke new ground or were innovative in some way. this could serve as a segue into a discussion of the repository as one more oppor- tunity to be innovative. once the self-concept has been boosted, the threatening quality of the message will be perceived as less disturbing and will be more likely to receive consideration. a related strategy that psychologists employ to reduce resistance involves casting the user in the role of “expert.” this is especially easy to do with scholars because they are experts in their fields. casting the user in the role of expert can deactivate resistance by putting that person in the persuasive role, which creates a form of role reversal. rather than the librarian being seen as the persuader, the scholar is placed in that role. by saying to the scholar, “you are the expert in the area of communicating your research to an audience, so you would know better why the digital repository is an alternative that deserves con- sideration once you understand how it works and how it may benefit you,” you are empowering the user. casting the user as an expert imparts a sense of control to the user. it helps to disable resistance by placing the user in a posi- tion of being predisposed to agree to the role he or she is being cast in, which also makes the user more prone to agree with the idea of using a digital repository. priming and imaging one important discovery that psychologists have made that has some bearing on user resistance is that even subtle manipulations can have a significant effect on one’s judgments and actions. in an interesting experiment, psy- chologists told a group of students that they were to read an online newspaper, ostensibly to evaluate its design and assess how easy it was to read. half of them read an editorial discussing a public opinion survey of youth ■■ negative and positive strategies for reducing resistance just as the definition of resistance can be paradoxical, so too may be some of the strategies that psychologists use to address it. perhaps the most basic example is to coun- ter resistance by acknowledging it. when scholars are presented with a message that overtly states that digital repositories are beneficial and desirable, it may simultane- ously generate a covert reaction in the form of resistance. rather than simply anticipating this and attempting to ignore it, digital repository advocates might be more persuasive if they acknowledge to scholars that there will likely be resistance, mention some possible reasons (e.g., plagiarism or copyright concerns), and immediately intro- duce some counterrationales to address those reasons. psychologists have found that being up front and forthcoming can reduce resistance, particularly with regard to the downside of digital repositories. they have learned that it can be advantageous to preemptively reveal negative information about something so that it can be downplayed or discounted. thus talking about the weak- nesses or shortcomings of digital repositories as early as possible in an interaction may have the effect of making these problems seem less important and weakening user resistance. not only does revealing negative information impart a sense of honesty and credibility to the user, but psychologists have found that people feel closer to people who reveal personal information. a librarian could thus describe some of his or her own frustrations in using repositories as an effective way of establishing rapport with resistant users. the unexpected approach of bring- ing up the less desirable aspects of repositories—whether this refers to the technological steps that must be learned to submit one’s work or the fact that depositing one’s work in a repository is not a guarantee that it will be highly cited—can be disarming to the resistant user. this is particularly true of more resistant users who may have been expecting a strong hard-sell approach on the part of librarians. when suddenly faced with a more candid appeal the user may be thrown off balance psychologi- cally, leaving him or her more vulnerable to information that is the opposite of what was anticipated and to pos- sibly viewing that information in a more positive light. if one way to disarm a user is to begin by discuss- ing the negatives, a seemingly opposite approach that psychologists take is to reinforce the user’s sense of self. psychologists believe that one source of resistance stems from when a user’s self-concept—which the user tries to protect from any source of undesired change—has been threatened in one way or another. a stable self-concept is necessary for the user to maintain a sense of order and predictability. reinforcing the self-concept of the user should therefore make the user less likely to resist depos- iting work in a digital repository. self-affirmed users are information technology and libraries | june or even possibly collaborating on research. their imagina- tions could be further stimulated by asking them to think of what it would be like to have their work still actively preserved and available to their successors a century from now. using the imagining strategy could potentially be significantly more effective in attenuating resistance than presenting arguments based on dry facts. identification and liking conscious processes like imagining are not the only psy- chological means of reducing the resistance of users to digital repositories. unconscious processes can also be helpful. one example of such a process is what psycholo- gists refer to as the “liking heuristic.” this refers to the tendency of users to employ a rule-of-thumb method to decide whether to comply with requests from persons. this tendency results from users constantly being inun- dated with requests. consequently, they need to simplify and streamline the decision-making process that they use to decide whether to cooperate with a request. the liking heuristic holds that users are more likely to help some- one they might otherwise not help if they unconsciously identify with the person. at an unconscious level, the user may think that a person acts like them and dresses like them, and therefore the user identifies with that person and likes them enough to comply with their request. in one experiment that psychologists conducted to see if people are more likely to comply with requests from people that they identify with, female undergraduates were informed that they would be participating in a study of first impressions. the subjects were instructed that they and a person in another room would each learn a little about one another without meeting each other. each sub- ject was then given a list of fifty adjectives and was asked to select the twenty that were most characteristic of them- selves. the experimenter then told the participants that they would get to see each other’s lists. the experimenter took the subject’s list and then returned a short time later with what supposedly was the other participant’s list, but was actually a list that the experimenter had filled out to indicate that either the subject had much in common with the other participant’s personality (seventeen of twenty matches), some shared attributes (ten of twenty matches), or relatively few characteristics in common (three of twenty matches). the subject was then asked to exam- ine the list and fill out a survey that probed their initial impressions of the other participant, including how much they liked them. at the end of the experiment, the two subjects were brought together and given credit for par- ticipating. the experimenter soon left the room and the confederate participant asked the other participant if she would read and critically evaluate an eight-page paper for an english class. the results of the experiment indi- cated that the more the participant thought she shared in consumer patterns that highlighted functional needs, and the other half read a similar editorial focusing on hedo- nistic needs. the students next viewed an ad for a new brand of shampoo that featured either a strong or a weak argument for the product. the results of the experiment indicated that students who read the functional editorial and were then subsequently exposed to the strong argu- ment for the shampoo (a functional product) had a much more favorable impression of the brand than students who had received the mismatched prime. while it may seem that the editorial and the shampoo were unrelated, psychologists found that the subjects engaged in a process of elaborating the editorial, which then predisposed them to favor the shampoo. the presence of elaboration, which is a precursor to the development of attitudes, suggests that librarians could reduce users’ resistance to digital repositories by first involving them in some form of priming activity immediately prior to any attempt to persuade them. for example, asking faculty to read a brief case study of a scholar who has benefited from involvement in open-access activity might serve as an effective prime. another example might be to listen briefly to a speaker summarizing the individual, disciplinary, and societal benefits of sharing one’s research with colleagues. interventions like these should help mitigate any predispo- sition toward resistance on the part of users. imagining is a strategy related to priming that psy- chologists have found to be effective in reducing resistance. taking their cue from insurance salesmen—who are trained to get clients to actively imagine what it would be like to lose their home or be in an accident—a group of psycholo- gists conducted an experiment in which they divided a sample of homeowners who were considering the purchase of cable tv into two groups. one group was presented with the benefits of cable in a straightforward, informative way that described various features. the other group was asked to imagine themselves enjoying the benefits and all the possible channels and shows that they might experi- ence and how entertaining it might be. the psychologists then administered a questionnaire. the results indicated that those participants who were asked to imagine the benefits of cable were much more likely to want cable tv and to subscribe to it than were those who were only given information about cable tv. in other words, imagining resulted in more positive attitudes and beliefs. this study suggests that librarians attempting to reduce resistance among users of digital repositories may need to do more than merely inform or describe to them the advan- tages of depositing their work. they may need to ask users to imagine in vivid detail what it would be like to receive periodic reports indicating that their work had been down- loaded dozens or even hundreds of times. librarians could ask them to imagine receiving e-mail or calls from col- leagues indicating that they had accessed their work in the repository and were interested in learning more about it, reducing psychological resistance to digital repositories | quinn students typically overestimate the amount of drinking that their peers engage in at parties. these inaccurate nor- mative beliefs act as a negative influence, causing them to imbibe more because they believe that is what their peers are doing. by informing students that almost three- quarters of their peers have less than three drinks at social gatherings, psychologists have had some success in reduc- ing excessive drinking behavior by students. the power of normative messages is illustrated by a recent experiment conducted by a group of psycholo- gists who created a series of five cards to encourage hotel guests to reuse their towels during their stay. the psychologists hypothesized that by appealing to social norms, they could increase compliance rates. to test their hypothesis, the researchers used a different conceptual appeal for each of the five cards. one card appealed to environmental concerns (“help save the environment”), another to environmental cooperation (“partner with us to save the environment”), a third card appealed to the advantage to the hotel (“help the hotel save energy”), a fourth card targeted future generations (“help save resources for future generations”), and a final card appealed to guests by making reference to a descrip- tive norm of the situation (“join your fellow citizens in helping to save the environment”). the results of the study indicated that the card that mentioned the benefit to the hotel was least effective in getting guests to reuse their towels, and the card that was most effective was the one that mentioned that descriptive norm. this research suggests that if users who are resistant to submitting their work to digital repositories were informed that a larger percentage of their peers were depositing work than they realized, resistance may be reduced. this might prove to be particularly true if they learned that prominent or influential scholars were engaged in popu- lating repositories with their work. this would create a social-norms effect that would help legitimize repositories to other faculty and help them to perceive the submission process as normal and desirable. the idea that accom- plished researchers are submitting materials and reaping the benefits might prove very attractive to less experienced and less well-regarded faculty. psychologists have a considerable body of evidence in the area of social modeling that suggests that people will imitate the behavior of others in social situations because that behavior provides an implicit guideline of what to do in a similar situation. a related finding is that the more influential people are, the more likely it is for others to emulate their actions. this is even more probable for high- status individuals who are skilled and attractive and who are capable of communicating what needs to be done to potential followers. social modeling addresses both the cognitive dimension of how resistant users should behave and also the affective dimension by offering models that serve as a source of motivation to resistant users to change common with the confederate, the more she liked her. the more she liked the confederate and experienced a percep- tion of consensus, the more likely she was to comply with her request to critique the paper. thus, when trying to overcome the resistance of users to depositing their work in a digital repository, it might make sense to consider who it is that is making the request. universities sometimes host scholarly communi- cation symposia that are not only aimed at getting faculty interested in open-access issues, but to urge them to sub- mit their work to the institution’s repositories. frequently, speakers at these symposia consist of academic administra- tors, members of scholarly communication or open-access advocacy organizations, or individuals in the library field. the research conducted by psychologists, however, sug- gests that appeals to scholars and researchers would be more effective if they were made by other scholars and those who are actively engaged in research. faculty are much more likely to identify with and cooperate with requests from their own tribe, as it were, and efforts need to be concentrated on getting faculty who are involved in and understand the value of repositories to articulate this to their colleagues. researchers who can personally testify to the benefits of depositing their work are most likely to be effective at convincing other researchers of the value of doing likewise and will be more effective at reducing resis- tance. librarians need to recognize who their potentially most effective spokespersons and advocates are, which the psychological research seems to suggest is faculty talking to other faculty. perceived consensus and social modeling the processes of faculty identification with peers and perceived consensus mentioned above can be further enhanced by informing researchers that other scholars are submitting their work, rather than merely telling research- ers why they should submit their work. information about the practices of others may help change beliefs because of the need to identify with other in-group members. this is particularly true of faculty, who are prone to making con- tinuous comparisons with their peers at other institutions and who are highly competitive by nature. once they are informed of the career advantages of depositing their work (in terms of professional visibility, collaboration opportuni- ties, etc.), and they are informed that other researchers have these advantages, this then becomes an impetus for them to submit their work to keep up with their peers and stay competitive. a perception of consensus is thus fostered—a feeling that if one’s peers are already depositing their work, this is a practice that one can more easily agree to. psychologists have leveraged the power of identifi- cation by using social-norms research to inform people about the reality of what constitutes normative behavior as opposed to people’s perceptions of it. for example, college information technology and libraries | june highly resistant users that may be unwilling to submit their work to a repository. rather than trying to prepare a strong argument based on reason and logic, psychologists believe that using a narrative approach may be more effective. this means conveying the facts about open access and digital repositories in the form of a story. stories are less rhetori- cal and tend not to be viewed by listeners as attempts at persuasion. the intent of the communicator and the coun- terresistant message are not as overt, and the intent of the message might not be obvious until it has already had a chance to influence the listener. a well-crafted narrative may be able to get under the radar of the listener before the listener has a chance to react defensively and revert to a mode of resistance. in a narrative, beliefs are rarely stated overtly but are implied, and implied beliefs are more diffi- cult to refute than overtly stated beliefs. listening to a story and wondering how it will turn out tends to use up much of the cognitive attentional capacity that might otherwise be devoted to counterarguing, which is another reason why using a narrative approach may be particularly effec- tive with users who are strongly resistant. the longer and more subtle nature of narratives may also make them less a target of resistance than more direct arguments. using a narrative approach, the case for submitting work to a repository might be presented not as a collection of dry facts or statistics, but rather as a story. the pro- tagonists are the researchers, and their struggle is to obtain recognition for their work and to advance scholarship by providing maximum access to the greatest audience of scholars and to obtain as much access as possible to the work of their peers so that they can build on it. the pro- tagonists are thwarted in their attempts to achieve their ends by avaricious publishers who obtain the work of researchers for free and then sell it back to them in the form of journal and database subscriptions and books for exor- bitant prices. these prices far exceed the rate of inflation or the budgets of universities to pay for them. the publishers engage in a series of mergers and acquisitions that swal- low up small publishing firms and result in the scholarly publishing enterprise being controlled by a few giant firms that offer unreasonable terms to users and make unreason- able demands when negotiating with them. presented in this dramatic way, the significance of scholar participation in digital repositories becomes magnified to an extent that it becomes more difficult to resist what may almost seem like an epic struggle between good and evil. and while this may be a greatly oversimplified example, it nonetheless provides a sense of the potential power of using a narrative approach as a technique to reduce resistance. introducing a time element into the attempt to per- suade users to deposit their work in digital repositories can play an important role in reducing resistance. given that faculty are highly competitive, introducing the idea not only that other faculty are submitting their work but that they are already benefiting as a result makes the their behavior in the desired direction. redefinition, consistency, and depersonalization another strategy that psychologists use to reduce resis- tance among users is to change the definition of the situation. resistant users see the process of submitting their research to the repository as an imposition at best. in their view, the last thing that they need is another obliga- tion or responsibility to burden their already busy lives. psychologists have learned that reframing a situation can reduce resistance by encouraging the user to look at the same phenomenon in a different way. in the current situ- ation, resistant users should be informed that depositing their work in a digital repository is not a burden but a way to raise their professional profile as researchers, to expose their work to a wider audience, and to heighten their visibility among not only their peers but a much larger potential audience that would be able to encounter their work on the web. seen in this way, the additional work of submission is less of a distraction and more of a career investment. moreover, this approach leverages a related psycho- logical concept that can be useful in helping to dissolve resistance. psychologists understand that inconsistency has a negative effect on self-esteem, so persuading users to believe that submitting their work to a digital repository is consistent with their past behavior can be motivating. the point needs to be emphasized with researchers that the act of submitting their work to a digital repository is not something strange and radical, but is consistent with prior actions intended to publicize and promote their work. a digital repository can be seen as analogous to a preprint, book, journal, or other tangible and familiar vehicles that faculty have used countless times to send their work out into the world. while the medium might have changed, the intention and the goal are the same. reframing the act of depositing as “old wine in new bottles” may help to undermine resistance. in approaching highly resistant individuals, psycholo- gists have discovered that it is essential to depersonalize any appeal to change their behavior. instead of saying, “you should reduce your caloric intake,” it is better to say, “it is important for people to reduce their caloric intake.” this helps to deflect and reduce the directive, judgmental, and prescriptive quality of the request, thus making it less likely to provoke resistance. suggestion can be much less threatening than prescription among users who may be suspicious and mistrusting. reverting to a third-per- son level of appeal may allow the message to get through without it being immediately rejected by the user. narrative, timing, and anticipation psychologists recommend another strategy to help defuse reducing psychological resistance to digital repositories | quinn technological platforms, and so on. this could be fol- lowed by a reminder to users that it is their choice—it is entirely up to them. this reminder that users have the freedom of choice may help to further counter any resis- tance generated as a result of instructions or inducements to anticipate regret. indeed, psychologists have found that reinstating a choice that was previously threatened can result in greater compliance than if the threat had never been introduced. offering users the freedom to choose between alterna- tives tends to make them more likely to comply. this is because having a choice enables users to both accept and resist the request rather than simply focus all their resis- tance on a single alternative. when presented with options, the user is able to satisfy the urge to resist by rejecting one option but is simultaneously motivated to accept another option; the user is aware that there are benefits to comply- ing and wants to take advantage of them but also wants to save face and not give in. by being offered several alterna- tives that nonetheless all commit to a similar outcome, the user is able to resist and accept at the same time. for example, one alternative option to self-archiving might be to present the faculty member with the option of an author- pays publishing model. the choice of alternatives allows the faculty member to be selective and discerning so that a sense of satisfaction is derived from the ability to resist by rejecting one alternative. at the same time, the librar- ian is able to gain compliance because one of the other alternatives that commits the faculty member to depositing research is accepted. options, comparisons, increments, and guarantees in addition to offering options, another way to erode user resistance to digital repositories is to use a comparative strategy. one technique is to first make a large request, such as “we would like you to submit all the articles that you have published in the last decade to the repository,” and then follow this with a more modest request, such as “we would appreciate it if you would please deposit all the articles you have published in the last year.” the origi- nal request becomes an “anchor” or point of reference in the mind of the user against which the subsequent request is then evaluated. setting a high anchor lessens user resis- tance by changing the user’s point of comparison of the second request from nothing (not depositing any work in the repository) to a higher value (submitting a decade of work). in this way, a high reference anchor is established for the second request, which makes it seem more reason- able in the newly created context of the higher value. the user is thus more likely to comply with the second request when it is framed in this way. using this comparative approach may also work because it creates a feeling of reciprocity in the user. when proposition much more salient. it not only suggests that submitting work is a process that results in a desirable outcome, but that the earlier one’s work is submitted, the more recognition will accrue and the more rapidly one’s career will advance. faculty may feel compelled to submit their work in an effort to remain competitive with their colleagues. one resource that may be par- ticularly helpful for working with skeptical faculty who want substantiation about the effect of self-archiving on scholarly impact is a bibliography created by the open citation project titled, “the effect of open access and downloads (hits) on citation impact: a bibliography of studies.” it provides substantial documentation of the effect that open access has on scholarly visibility. an additional stimulus might be introduced in conjunction with the time element in the form of a download report. showing faculty how downloads accumulate over time is analogous to arguments that investment counselors use showing how interest on investments accrues and compounds over time. this investment analogy creates a condition in which hesitating to submit their work results in faculty potentially losing recognition and compromis- ing their career advancement. an interesting related finding by psychologists sug- gests that an effective way to reduce user resistance is to have users think about the future consequences of complying or not complying. in particular, if users are asked to anticipate the amount of future regret they might experience for making a poor choice, this can significantly reduce the amount of resistance to complying with a request. normally, users tend not to ruminate about the possibility of future disappointment in making a decision. if users are made to anticipate future regret, however, they will act in the present to try to minimize it. studies conducted by psychologists show that when users are asked to anticipate the amount of future regret that they might experience for choosing to comply with a request and having it turn out adversely versus choosing to not comply and having it turn out adversely, they consis- tently indicate that they would feel more regret if they did not comply and experienced negative consequences as a result. in an effort to minimize this anticipated regret, they will then be more prone to comply. based on this research, one strategy to reduce user resistance to digital repositories would be to get users to think about the future, specifically about future regret resulting from not cooperating with the request to sub- mit their work. if they feel that they might experience more regret in not cooperating than in cooperating, they might then be more inclined to cooperate. getting users to think about the future could be done by asking users to imagine various scenarios involving the negative out- comes of not complying, such as lost opportunities for recognition, a lack of citation by peers, lost invitations to collaborate, an inability to migrate one’s work to future information technology and libraries | june submit their work. mandates rely on authority rather than persuasion to accomplish this and, as such, may represent a less-than-optimal solution to reducing user resistance. mandates represent a failure to arrive at a meeting of the minds of advocates of open access, such as librarians, and the rest of the intellectual community. understanding the psychology of resistance is an important prerequisite to any effort to reduce it. psychologists have assembled a significant body of research on resistance and how to address it. some of the strategies that the research suggests may be effective, such as discussing resistance itself with users and talk- ing about the negative effects of repositories, may seem counterintuitive and have probably not been widely used by librarians. yet when other more conventional tech- niques have been tried with little or no success, it may make sense to experiment with some of these approaches. particularly in the academy, where reason is supposed to prevail over authority, incorporating resistance psychol- ogy into a program aimed at soliciting faculty research seems an appropriate step before resorting to mandates. most strategies that librarians have used in trying to persuade faculty to submit their work have been con- ventional. they are primarily of a cognitive nature and are variations on informing and educating faculty about how repositories work and why they are important. researchers have an important affective dimension that needs to be addressed by these appeals, and the psy- chological research on resistance suggests that a strictly rational approach may not be sufficient. by incorporating some of the seemingly paradoxical and counterintuitive techniques discussed earlier, librarians may be able to penetrate the resistance of researchers and reach them at a deeper, less rational level. ideally, a mixture of rational and less-conventional approaches might be combined to maximize effectiveness. such a program may not elimi- nate resistance but could go a long way toward reducing it. future studies that test the effectiveness of such pro- grams will hopefully be conducted to provide us with a better sense of how they work in real-world settings. references . charles w. bailey jr., “institutional repositories: doa?,” online posting, digital koans, aug. , , http://digital -scholarship.org/digitalkoans/ / / /institutional -repositories-doa/ (accessed apr. , ). . dorothea salo, “yes, irs are broken. let’s talk about it,” online posting, caveat lector, sept. , , http://cavlec. yarinareth.net/ / / /yes-irs-are-broken-lets-talk-about -it/ (accessed apr. , ). . eprints services, roarmap (registry of open access repository material archiving policies) http://www.eprints .org/openaccess/policysignup/ (accessed july , ). . richard k. johnson, “institutional repositories: partnering the requester scales down the request from the large one to a smaller one, it creates a sense of obligation on the part of the user to also make a concession by agreeing to the more modest request. the cultural expectation of reciprocity places the user in a situation in which they will comply with the lesser request to avoid feelings of guilt. for the most resistant users, breaking the request down into the smallest possible increment may prove helpful. by making the request seem more manageable, the user is encouraged to comply. psychologists conducted an experi- ment to test whether minimizing a request would result in greater cooperation. they went door-to-door, soliciting contributions to the american cancer society, and received donations from percent of households. they then made additional solicitations, this time asking, “would you contribute? even a penny will help!” using this approach, donations increased to percent. even though the solici- tors only asked for a penny, the amounts of the donations were equal to that of the original request. by asking for “even a penny,” the solicitors made the request appear to be more modest and less of a target of resistance. librarians might approach faculty by saying “if you could even submit one paper we would be grateful,” with the idea that once faculty make an initial submission they will be more inclined to submit more papers in the future. one final strategy that psychological research sug- gests may be effective in reducing resistance to digital repositories is to make sure that users understand that the decision to deposit their work is not irrevocable. with any new product, users have fears about what might hap- pen if they try it and they are not satisfied with it. not knowing the consequences of making a decision that they may later regret fuels reluctance to become involved with it. faculty need to be reassured that they can opt out of participating at any time and that the repository sponsors will guarantee this. this guarantee needs to be repeated and emphasized as much as possible in the solicitation process so that faculty are frequently reminded that they are entering into a decision that they can reverse if they so decide. having this reassurance should make research- ers much less resistant to submitting their work, and the few faculty who may decide that they want to opt out are worth the reduction in resistance. the digital repository is a new phenomenon that faculty are unfamiliar with, and it is therefore important to create an atmosphere of trust. the guarantee will help win that trust. ■■ conclusion the scholarly literature on digital repositories has given little attention to the psychology of resistance. yet the ultimate success of digital repositories depends on over- coming the resistance of scholars and researchers to reducing psychological resistance to digital repositories | quinn . curtis p. haugtvedt et al., “consumer psychology and attitude change,” in knowles and linn, resistance and persua- sion, – . . larry w. gregory, robert b. cialdini, and kathleen m. carpenter, “self-relevant scenarios as mediators of likelihood estimates and compliance: does imagining make it so?” journal of personality & social psychology , no. ( ): – . . jerry m. burger, “fleeting attraction and compliance with requests,” in the science of social influence: advances and future progress, ed. anthony r. pratkanis (new york: psychol- ogy pr., ): – . . john d. clapp and anita lyn mcdonald, “the relation- ship of perceptions of alcohol promotion and peer drinking norms to alcohol problems reported by college students,” journal of college student development , no. ( ): – . . noah j. goldstein and robert b. cialdini, “using social norms as a lever of social influence,” in the science of social influence: advances and future progress, ed. anthony r. pratkanis (new york: psychology pr., ): – . . dale h. schunk, “social-self interaction and achievement behavior,” educational psychologist , no. ( ): – . . rosanna e. guadagno et al., “when saying yes leads to saying no: preference for consistency and the reverse foot-in- the-door effect,” personality & social psychology bulletin , no. ( ): – . . mary jiang bresnahan et al., “personal and cultural dif- ferences in responding to criticism in three countries,” asian journal of social psychology , no. ( ): – . . melanie c. green and timothy c. brock, “in the mind’s eye: transportation-imagery model of narrative persuasion,” in narrative impact: social and cultural foundations, ed. melanie c. green, jeffrey j. strange, and timothy c. brock (mahwah, n.j.: lawrence erlbaum, ): – . . oswald huber, “time pressure in risky decision mak- ing: effect on risk defusing,” psychology science , no. ( ): – . . the open citation project, “the effect of open access and downloads (‘hits’) on citation impact: a bibliography of studies,” july , , http://opcit.eprints.org/oacitation -biblio.html (accessed july , ). . matthew t. crawford et al., “reactance, compliance, and anticipated regret,” journal of experimental social psychology , no. ( ): – . . nicolas gueguen and alexandre pascual, “evocation of freedom and compliance: the ‘but you are free of . . .’ tech- nique,” current research in social psychology , no. ( ): – . . james p. dillard, “the current status of research on sequential request compliance techniques,” personality & social psychology bulletin , no. ( ): – . . thomas mussweiler, “the malleability of anchoring effects,” experimental psychology , no. ( ): – . . robert b. cialdini and noah j. goldstein, “social influ- ence: compliance and conformity,” annual review of psychology ( ): – . . james m. wyant and stephen l. smith, “getting more by asking for less: the effects of request size on donations of char- ity,” journal of applied social psychology , no. ( ): – . . lydia j. price, “the joint effects of brands and warranties in signaling new product quality,” journal of economic psychol- ogy , no. ( ): – . with faculty to enhance scholarly communication,” d-lib mag- azine , no. ( ), http://www.dlib.org/dlib/november / johnson/ johnson.html (accessed apr. , ). . bruce heterick, “faculty attitudes toward electronic resources,” educause review , no. ( ): – . . nancy fried foster and susan gibbons, “understanding faculty to improve content recruitment for institutional repos- itories,” d-lib magazine , no. ( ), http://www.dlib.org/ dlib/january /foster/ foster.html (accessed july , ). . suzanne bell, nancy fried foster, and susan gibbons, “reference librarians and the success of institutional reposito- ries,” reference services review , no. ( ): – . . diane harley et al., “the influence of academic values on scholarly publication and communication practices,” center for studies in higher education, research & occasional paper series: cshe. . , sept. , , http://repositories.cdlib.org/ cshe/cshe- - / (accessed apr. , ). . rea devakos, “towards user responsive institutional repositories: a case study,” library high tech , no. ( ): – . . philip m. davis and matthew j. l. connolly, “institutional repositories: evaluating the reasons for non-use of cornell university’s installation of dspace,” d-lib magazine , no. / ( ), http://www.dlib.org/dlib/march /davis/ davis .html (accessed july , ). . jihyun kim, “motivating and impeding factors affecting faculty contribution to institutional repositories,” journal of digital information , no. ( ), http://journals.tdl.org/jodi/ article/view/ / (accessed july , ). . peter suber, “open access overview” online posting, open access news: news from the open access environment, june , , http://www.earlham.edu/~peters/fos/overview .htm (accessed july ). . see, for example, jeffrey d. ford and laurie w. ford, “decoding resistance to change,” harvard business review , no. ( ): – .; john p. kotter and leonard a. schlesinger, “choosing strategies for change,” harvard business review , no. / ( ): – ; and paul r. lawrence, “how to deal with resistance to change,” harvard business review , no. ( ): – . . julia zuwerink jacks and maureen e. o’brien, “decreas- ing resistance by affirming the self,” in resistance and per- suasion, ed. eric s. knowles and jay a. linn (mahwah, n.j.: lawrence erlbaum, ): – . . benjamin margolis, “notes on narcissistic resistance,” modern psychoanalysis , no. ( ): – . . ralph grabhorn et al., “the therapeutic relationship as reflected in linguistic interaction: work on resistance,” psycho- therapy research , no. ( ): – . . arthur aron et al., “the experimental generation of interpersonal closeness: a procedure and some preliminary findings,” personality & social psychology bulletin , no. ( ): – . . geoffrey l. cohen, joshua aronson, and claude m. steele, “when beliefs yield to evidence: reducing biased evaluation by affirming the self,” personality & social psychology bulletin , no. ( ): – . . anthony r. pratkanis, “altercasting as an influence tac- tic,” in attitudes, behavior and social context: the role of norms and group membership, ed. deborah j. terry and michael a.hogg (mahwah, n.j.: lawrence erlbaum, ): – . information technology and libraries | march michelle frisque (mfrisque@northwestern.edu) is lita president – and head, information systems, north- western university, chicago. michelle frisque michelle frisque (mfrisque@northwestern.edu) is lita president – and head, information systems, north- western university, chicago. michelle frisque president’s message: join us at the forum! t he first lita national forum i attended was in milwaukee, wisconsin. it seems like it was only a couple of years ago, but in fact nine national forums have since passed. i was a new librarian, and i went on a lark when a colleague invited me to attend and let me crash in her room for free. i am so glad i took her up on the offer because it was one of the best conferences i have ever attended. it was the first conference that i felt was made up of people like me, people who shared my interests in technology within the library. the program- ming was a good mix of practical know-how and mind- blowing possibilities. my understanding of what was possible was greatly expanded, and i came home excited and ready to try out the new things i had learned. almost eight years passed before i attended my next forum in cincinnati, ohio. after half a day i wondered why i had waited so long. the program was diverse, cov- ering a wide range of topics. i remember being depressed and outraged on the current state of internet access in the united states as reported by the office for information technology policy. i felt that surge of recognition when i discovered that other universities were having a difficult time documenting and tracking the various systems they run and maintain. i was inspired by david lanke’s talk, “obligations of leadership.” if you missed it you can still hear it online. it is linked from the lita blog (http:// www.litablog.org). while the next forum may seem like a long way off to you, it is in the forefront of my mind. the national forum planning committee is busy working to make sure this forum lives up to the reputation of forums past. this year’s forum takes place in atlanta, georgia, september –october . the theme is “the cloud and the crowd.” program proposals are due february , so i cannot give you specifics about the concurrent sessions, but we do hope to have presentations about projects, plans, or discoveries in areas of library-related technology involv- ing emerging cloud technologies; software-as-service, as well as social technologies of various kinds; using vir- tualized or cloud resources for storage or computing in libraries; library-specific open-source software (oss) and other oss “in” libraries; technology on a budget; using crowdsourcing and user groups for supporting technol- ogy projects; and training via the crowd. each accepted program is scheduled to maximize the impact for each attendee. programming ranges from five-minute lightening talks to full day preconferences. in addition, on the basis of attendee comments from previ- ous forums, we have also decided to offer thirty- and seventy-five-minute concurrent sessions. these concur- rent sessions will be a mix of traditional single- or multi- speaker formats, panel discussions, case studies, and demonstrations of projects. finally, poster sessions will also be available. while programs such as the keynote speakers, light- ning talks, and concurrent sessions are an important part of the forum experience, so is the opportunity to network with other attendees. i know i have learned just as much talking with a group of people in the hall between ses- sions, during lunch, or at the networking dinners as i have sitting in the programs. not only is it a great oppor- tunity to catch up with old friends, you will also have the opportunity to make new ones. for instance, at the national forum in salt lake city, utah, approximately half of the people who attended were first-time attendees. the national forum is an intimate event whose atten- dance ranges between and people, thus making it easy to forge personal connections. attendees come from a variety of settings, including academic, public, and spe- cial libraries; library-related organizations; and vendors. if you want to meet the attendees in a more formal setting you can attend a networking dinner organized on-site by lita members. this year the dinners were organized by the lita president, lita past president, lita president- elect, and a lita director-at-large. if you have not attended a national forum or it has been a while, i hope i have piqued your interest in com- ing to the next national forum in atlanta. registration will open in may! the most up-to-date information about the forum is available at the lita website (http:// www.lita.org). i know that even after my lita presidency is a distant memory, i will still make time to attend the lita national forum. i hope to see you there! information technology and libraries | june in this paper we discuss the design space of meth- ods for integrating information from web services into websites. we focus primarily on client-side mash-ups, in which code running in the user’s browser contacts web services directly without the assistance of an inter- mediary server or proxy. to create such mash-ups, we advocate the use of “widgets,” which are easy-to-use, customizable html elements whose use does not require programming knowledge. although the techniques we discuss apply to any web-based information system, we specifically consider how an opac can become both the target of web services integration and also a web service that provides information to be integrated elsewhere. we describe three widget libraries we have developed, which provide access to four web services. these libraries have been deployed by us and others. our contributions are twofold: we give practitioners an insight into the trade-offs surrounding the appropri- ate choice of mash-up model, and we present the specific designs and use examples of three concrete widget libraries librarians can directly use or adapt. all software described in this paper is available under the lgpl open source license. ■■ background web-based information systems use a client-server archi- tecture in which the server sends html markup to the user’s browser, which then renders this html and dis- plays it to the user. along with html markup, a server may send javascript code that executes in the user’s browser. this javascript code can in turn contact the original server or additional servers and include infor- mation obtained from them into the rendered content while it is being displayed. this basic architecture allows for myriad possible design choices and combinations for mash-ups. each design choice has implications to ease of use, customizability, programming requirements, hosting requirements, scalability, latency, and availability. server-side mash-ups in a server-side mash-up design, shown in figure , the mash-up server contacts the base server and each source when it receives a request from a client. it combines the information received from the base server and the sources and sends the combined html to the client. server-side mash-up systems that combine base and mash-up servers are also referred to as data mash-up systems. such data mash-up systems typically provide a web-based configuration front-end that allows users to select data sources, specify the manner in which they are combined, and to create a layout for the entire mash-up. godmar back and annette bailey web services and widgets for library information systems as more libraries integrate information from web services to enhance their online public displays, techniques that facilitate this integration are needed. this paper presents a technique for such integration that is based on html widgets. we discuss three example systems (google book classes, tictoclookup, and majax) that implement this technique. these systems can be easily adapted without requiring programming experience or expensive hosting. t o improve the usefulness and quality of their online public access catalogs (opacs), more and more librarians include information from addi- tional sources into their public displays. examples of such sources include web services that provide addi- tional bibliographic information, social bookmarking and tagging information, book reviews, alternative sources for bibliographic items, table-of-contents previews, and excerpts. as new web services emerge, librarians quickly integrate them to enhance the quality of their opac displays. conversely, librarians are interested in opening the bibliographic, holdings, and circulation information contained in their opacs for inclusion into other web offerings they or others maintain. for example, by turn- ing their opac into a web service, subject librarians can include up-to-the-minute circulation information in sub- ject or resource guides. similarly, university instructors can use an opac’s metadata records to display citation information ready for import into citation management software on their course pages. the ability to easily create such “mash-up” pages is crucial for increasing the vis- ibility and reach of the digital resources libraries provide. although the technology to use web services to create mash-ups is well known, several practical requirements must be met to facilitate its widespread use. first, any environment providing for such integration should be easy to use, even for librarians with limited programming background. this ease of use must extend to environments that include proprietary systems, such as vendor-provided opacs. second, integration must be seamless and custom- izable, allowing for local display preferences and flexible styling. third, the setup, hosting, and maintenance of any necessary infrastructure must be low-cost and should maximize the use of already available or freely accessible resources. fourth, performance must be acceptable, both in terms of latency and scalability. godmar back (gback@cs.vt.edu) is assistant professor, depart- ment of computer science and annette bailey (afbailey@vt.edu) is assistant professor, university libraries, virginia tech univer- sity, blacksburg. web services and widgets for library information systems | back and bailey examples of such systems include dapper and yahoo! pipes. these systems require very little programming knowledge, but they limit mash-up creators to the func- tionality supported by a particular system and do not allow the user to leverage the layout and functionality of an existing base server, such as an existing opac. integrating server-side mash-up systems with pro- prietary opacs as the base server is difficult because the mash-up server must parse the opac’s output before integrating any additional information. moreover, users must now visit—or be redirected to—the url of the mash-up server. although some emerging extensible opac designs provide the ability to include information from external sources directly and easily, most currently deployed systems do not. in addition, those mash-up servers that do usually require server-side programming to retrieve and integrate the information coming from the mash-up sources into the page. the availability of software libraries and the use of special purpose markup languages may mitigate this requirement in the future. from a performance scalability point of view, the mash-up server is a bottleneck in server-side mash-ups and therefore must be made large enough to handle the expected load of end-user requests. on the other hand, the caching of data retrieved from mash-up sources is simple to implement in this arrangement because only the mash-up server contacts these sources. such caching reduces the frequency with which requests have to be sent to sources if their data is cacheable, that is, if real- time information is not required. the latency in this design is the sum of the time required for the client to send a request to the mash- up server and receive a reply, plus the processing time required by the server, plus the time incurred by sending a request and receiving a reply from the last responding mash-up source. this model assumes that the mash-up server contacts all sources in parallel, or as soon as the server knows that information from a source should be included in a page. the availability of the system depends on the avail- ability of all mash-up sources. if a mash-up source does not respond, the end user must wait until such failure is apparent to the mash-up server via a timeout. finally, because the mash-up server acts as a client to the base and source servers, no additional security considerations apply with respect to which sources may be contacted. there also are no restrictions on the data interchange for- mat used by source servers as long as the mash-up server is able to parse the data returned. client-side mash-ups in a client-side setup, shown in figure , the base server sends only a partial website to the client, along with javascript code that instructs the client which other sources of information to contact. when executed in the browser, this javascript code retrieves the information from the mash-up sources directly and completes the mash-up. the primary appeal of client-side mashing is that no mash-up server is required, and thus the url that users visit does not change. consequently, the mash-up server is no longer a bottleneck. equally important, no main- tenance is required for this server, which is particularly relevant when libraries use turnkey solutions that restrict administrative access to the machine housing their opac. on the other hand, without a mash-up server, results from mash-up sources can no longer be centrally cached. thus the mash-up sources themselves must be sufficiently figure . server-side mash-up construction figure . client-side mash-up construction information technology and libraries | june scalable to handle the expected number of requests. as a load-reducing strategy, mash-up sources can label their results with appropriate expiration times to influence the caching of results in the clients’ browsers. availability is increased because the mash-up degrades gracefully if some of the mash-up sources fail, since the information from the remaining sources can still be dis- played to the user. assuming that requests are sent by the client in parallel or as soon as possible, and assuming that each mash-up source responds with similar latency to requests sent by the user’s browser as to requests sent by a mash-up server, the latency for a client-side mash-up is similar to the server-side mash-up. however, unlike in the server-side approach, the page designer has the option to display partial results to the user while some requests are still in progress, or even to delay sending some requests until the user explicitly requests the data by clicking on a link or other element on the page. because client-side mash-ups rely on javascript code to contact web services directly, they are subject to a number of restrictions that stem from the security model governing the execution of javascript code in current browsers. this security model is designed to protect the user from malicious websites that could exploit client-side code and abuse the user’s credentials to retrieve html or xml data from other websites to which a user has access. such malicious code could then relay this potentially sensitive data back to the malicious site. to prevent such attacks, the security model allows the retrieval of html text or xml data only from sites within the same domain as the origin site, a policy commonly known as same- origin policy. in figure , sources a and b come from the same domain as the page the user visits. the restrictions of the same-origin policy can be avoided by using the javascript object notation (json) interchange format. because client-side code may retrieve and execute javascript code served from any domain, web services that are not co-located with the origin site can make their results available using json. doing so facilitates their inclusion into any page, independent of the domain from which it is served (see source c in figure ). many existing web services already provide an option to return data in json format, perhaps along with other formats such as xml. for web services that do not, a proxy server may be required to translate the data com- ing from the service into json. if the implementation of a proxy server is not feasible, the web service is usable only on pages within the same domain as the website using it. client-side mash-ups lend themselves naturally to enhancing the functionality of existing, proprietary opac systems, particularly when a vendor provides only lim- ited extensibility. because they do not require server-side programming, the absence of a suitable vendor-provided server-side programming interface does not prevent their creation. oftentimes, vendor-provided templates or variables can be suitably adapted to send the necessary html markup and javascript code to the client. the amount of javascript code a librarian needs to write (or copy from a provided example) determines both the likelihood of adoption and the maintainability of a given mash-up creation. the less javascript code there is to write, the larger the group of librarians who feel comfortable try- ing and adopting a given implementation. the approach of using html widgets hides the use of javascript almost entirely from the mash-up creator. html widgets repre- sent specially composed markup, which will be replaced with information coming from a mash-up source when the page is rendered. because the necessary code is contained in a javascript library, adapters do not need to understand programming to use the information coming from the web service. finally, html widgets are also preferable for javascript-savvy users because they create a layer of abstraction over the complexity and browser dependencies inherent in javascript programming. ■■ the google book classes widget library to illustrate our approach, we present a first example that allows the integration of data obtained from google book search into any website, including opac pages. google book search provides access to google’s database of book metadata and contents. because of the company’s book scanning activities as well as through agreements with publishers, google hosts scanned images of many book jackets as well as partial or even full previews for some books. many libraries are interested in either using the book jackets when displaying opac records or alerting their users if google can provide a partial or full view of an item a user selected in their catalog, or both. this service can help users decide whether to borrow the book from the library. the google book search dynamic link api the google book search dynamic link api is a json- based web service through which google provides certain metadata for items it has indexed. it can be queried using bibliographic identifiers such as isbn, oclc number, or library of congress control number (lccn). it returns a small set of data that includes the url of a book jacket thumbnail image, the url of a page with bibliographic information, the url of a preview page (if available), as well as information about the extent of any preview and whether the preview viewer can be embedded directly into other pages. table shows the json result returned for an example isbn. web services and widgets for library information systems | back and bailey widgetization to facilitate the easy integration of this service into web- sites without javascript programming, we developed a widget library. from the adapter’s perspective, the use of these widgets is extremely simple. the adapter places html or
tags into the page where they want data from google book search to display. these tags contain an html attribute that acts as an identifier to describe the bibliographic item for which information should be retrieved. it may contain its isbn, oclc num- ber, or lccn. in addition, the tags also contain one or more html <class> attributes to describe which processing should be done with the information retrieved from google to integrate it into the page. these classes can be combined with a list of traditional css classes in the <class> attribute to apply further style and formatting control. examples as an example, consider the following html an adapter may use in a page: <span title=“isbn: ” class=“gbs -thumbnail gbs-link-to-preview”></span> when processed by the google book classes widget library, the class “gbs-thumbnail” instructs the widget to embed a thumbnail image of the book jacket for isbn , and “gbs-link-to-preview” provides instruc- tions to wrap the <span> tag in a hyperlink pointing to google’s preview page. the result is as if the server had contacted google’s web service and constructed the html shown in example in table , but the mash-up creator does not need to be concerned with the mechanics of contacting google’s service and making the necessary manipulations to the document. example in table demonstrates a second possible use of the widget. in this example, the creator’s intent is to display an image that links to google’s information page if and only if google provides at least a partial preview for the book in question. this goal is accom- plished by placing the image inside the span and using style=“display:none” to make the span initially invisible. the span is made visible only if a preview is available at google, displaying the hyperlinked image. the full list of features supported by the google book classes widget library can be found in table . integration with legacy opacs the approach described thus far assumes that the mash- up creator has sufficient control over the html markup that is sent to the user. this assumption does not always hold if the html is produced by a vendor-provided system, since such systems automatically generate most of the html used to display opac search results or indi- vidual bibliographic records. if the opac provides an extension system, such as a facility to embed customized links to external resources, it may be used to generate the necessary html by utilizing variables (e.g., “@#isbn@” for isbn numbers) set by the opac software. if no extension facility exists, accommodations by the widget library are needed to maintain the goal of not requiring any programming on the part of the adapter. we implemented such accommodations to facilitate the use of google book classes within a iii millennium opac. we used magic strings such as “isbn:millennium.record” in a table . sample request and response for google book search dynamic link api request: http://books.google.com/books?bibkeys=isbn: &jscmd=viewapi&callback=process json response: process({ “isbn: ”: { “bib_key”: “isbn: ”, “info_url”: “http://books.google.com/books?id=ezqe hh q c\x source=gbs_viewapi”, “preview_url”: “http://books.google.com/books?id=ezqe hh q c\x printsec=frontcover\x source=gbs_viewapi”, “thumbnail_url”: “http://bks .books.google.com/books?id=ezqe hh q c\x printsec=frontcover\x img= \x zoom= \x sig=acfu u d usnxw baqd u nc quwhjn a”, “preview”: “partial”, “embeddable”: true } }); information technology and libraries | june table . example of client-side processing by the google book classes widget library example : html written by adapter browser display <span title=“isbn: ” class=“gbs-thumbnail gbs-link-to-preview”> </span> resultant html after client-side processing <a href=“http://books.google.com/books?id=ezqe hh q c& printsec=frontcover&source=gbs_viewapi”> <span title=“” class=”gbs-thumbnail gbs-link-to-preview”> <img src=“http://bks .books.google.com/books?id=ezqe hh q c& amp;printsec=frontcover&img= &zoom= & sig=acfu u d usnxw baqd u nc quwhjn a” /> </span> </a> example : html written by adapter browser display <span style=“display: none” title=“isbn: ” class=“gbs-link-to-info gbs-if-partial-or-full”> <img src=“http://www.google.com/intl/en/googlebooks/images/ gbs_preview_button .gif” /> </span> resultant html after client-side processing <a href=”http://books.google.com/books?id=ezqe hh q c& source=gbs_viewapi”> <span title=“” class=“gbs-link-to-info gbs-if-partial-or-full”> <img src=“http://www.google.com/intl/en/googlebooks/images/ gbs_preview_button .gif” /> </span> </a> table . supported google book classes google book class meaning gbs-thumbnail gbs-link-to-preview gbs-link-to-info gbs-link-to-thumbnail gbs-embed-viewer gbs-if-noview gbs-if-partial-or-full gbs-if-partial gbs-if-full gbs-remove-on-failure include an <img...> embedding the thumbnail image wrap span/div in link to preview at google book search (gbs) wrap span/div in link to info page at gbs wrap span/div in link to thumbnail at gbs directly embed a viewer for book’s content into the page, if possible keep this span/div only if gbs reports that book’s viewability is “noview” keep this span/div only if gbs reports that book’s viewability is at least “partial” keep this span/div only if gbs reports that book’s viewability is “partial” keep this span/div only if gbs reports that book’s viewability is “full” remove this span/div if gbs doesn’t return book information for this item <title> attribute to instruct the widget library to harvest the isbn from the current page via screen scraping. figure provides an example of how a google book classes widget can be integrated into an opac search results page. ■■ the tictoclookup widget library the tictocs journal table of contents service is a free online service that allows academic researchers and web services and widgets for library information systems | back and bailey other users to keep up with newly published research by giving them access to thousands of journal tables of con- tents from multiple publishers. the tictocs consortium compiles and maintains a dataset that maps issns and journal titles to rss-feed urls for the journals’ tables of contents. the tictoclookup web service we used the tictocs dataset to create a simple json web service called “tictoclookup” that returns rss-feed urls when queried by issn and, optionally, by journal title. table shows an example query and response. to accommodate different hosting scenarios, we created two implementations of this tictoclookup: a standalone and a cloud-based implementation. the standalone version is implemented as a python web application conformant to the web services gateway interface (wsgi) specification. hosting this version requires access to a web server that supports a wsgi- compatible environment, such as apache’s mod_wsgi. the python application reads the tictocs dataset and responds to lookup requests for specific issns. a cron job downloads the most up-to-date version of the dataset periodically. the cloud version of the tictoclookup service is implemented as a google app engine (gae) applica- tion. it uses the highly scalable and highly available gae datastore to store tictocs data records. gae applications run on servers located in google’s regional data centers so that requests are handled by a data center geographically close to the requesting client. as of june , google hosting of gae applications is free, which includes a free allotment of several computational resources. for each application, gae allows quotas of up to . mb requests and the use of up to gb of bandwidth per twenty-four- hour period. although this capacity is sufficient for the purposes of many small- and medium-size institutions, additional capacity can be purchased at a small cost. widgetization to facilitate the easy integration of this service into websites without javascript programming, we developed a widget library. like google book classes, this widget library is controlled via html attributes associated with html <span> or <div> tags that are placed into the page where the user decides to display data from the tictoclookup service. the html <title> attribute identifies the journal by its issn or its issn and title. as with google book classes, figure . sample use of google book classes in an opac results page table . sample request and response for tictocs lookup web service request: http://tictoclookup.appspot.com/ - ?title=nature&jsoncallback=process json response: process({ “lastmod”: “wed apr : : ”, “records”: [{ “title”: “nature”, “rssfeed”: http://www.nature.com/nature/current_issue/rss }], “issn”: “ ” }); information technology and libraries | june the html <class> attribute describes the desired process- ing, which may contain traditional css classes. example consider the following html an adapter may use in a page: <span style=“display:none” class=“tictoc-link tictoc-preview tictoc-alternate-link” title=“issn: : nature”> click to subscribe to table of contents for this journal </span> when processed by the tictoclookup widget library, the class “tictoc-link” instructs the widget to wrap the span in a link to the rss feed at which the table of con- tent is published, allowing users to subscribe to it. the class “tictoc-preview” associates a tooltip element with the span, which displays the first entries of the feed when the user hovers over the link. we use the google feeds api, another json-based web service, to retrieve a cached copy of the feed. the “tictoc-alternate-link” class places an alternate link into the current document, which in some browsers triggers the display of the rss feed icon figure . sample use of tictoclookup classes in the status bar. the <span> element, which is initially invisible, is made visible if and only if the tictoclookup service returns information for the given pair of issn and title. figure provides a screenshot of the display if the user hovers over the link. as with google book classes, the mash-up creator does not need to be concerned with the mechanics of contacting the tictoclookup web service and making the necessary manipulations to the document. table provides a com- plete overview of the classes tictoclookup supports. integration with legacy opacs similar to the google book classes widget library, we implemented provisions that allow the use of tictoclookup classes on pages over which the mash-up creator has limited control. for instance, specifying a title attribute of “issn:millennium.issnandtitle” harvests the issn and journal title from the iii millennium’s record display page. ■■ majax whereas the widget libraries discussed thus far integrate external web services into an opac display, majax is a widget library that integrates information coming from an opac into other pages, such as resource guides or course displays. majax is designed for use with a iii millennium integrated library system (ils) whose vendor does not provide a web-services interface. the tech- niques we used, however, extend to other opacs as well. like many table . supported tictoclookup classes tictoclookup class meaning tictoc-link tictoc-preview tictoc-embed-n tictoc-alternate-link tictoc-append-title wrap span/div in link to table of contents display tooltip with preview of current entries embed preview of first n entries insert <link rel=“alternate”> into document append the title of the journal to the span/div web services and widgets for library information systems | back and bailey legacy opacs, millennium does not only lack a web-services interface, but lacks any programming interface to the records contained in the system and does not provide access to the database or file system of the machine housing the opac. providing opac data as a web service we implemented two methods to access records from the millennium opac using bibliographic identifi- ers such as isbn, oclc number, bibliographic record number, and item title. both methods provide access to complete marc records and holdings information, along with locations and real-time availability for each held item. majax extracts this information via screen- scraping from the marc record display page. as with all screen-scraping approaches, the code performing the scraping must be updated if the output format provided by the opac changes. in our experience, such changes occur at a frequency of less than once per year. the first method, majax , implements screen scrap- ing using javascript code that is contained in a document placed in a directory on the server (/screens), which is normally used for supplementary resources, such as images. this document is included in the target page as a hidden html <iframe> element (see frame b in figure ). consequently, the same-domain restriction applies to the code residing in it. majax can thus be used only on pages within the same domain—for instance, if the opac is housed at opac.library.university.edu, majax may be used on all pages within *.university.edu (not merely *.library.university.edu). the key advantage of majax is that no additional server is required. the second method, majax , uses an intermediary server that retrieves the data from the opac, translates it to json, and returns it to the client. this method, shown in figure , returns json data and therefore does not suffer from the same-domain restriction. however, it requires hosting the majax web service. like the tictoclookup web service, we implemented the majax web service using python conformant to wsgi. a single installation can support multiple opacs. widgetization the majax widget library allows the integration of both majax and majax data into websites without javascript programming. the <span> tags function as placeholders, and <title> and <class> attributes describe the desired processing. majax provides a number of “majax classes,” multiple of which can be specified. these classes allow a mash-up creator to insert a large variety of bibliographic information, such as the val- ues of marc fields. classes are also provided to insert fully formatted, ready-to-copy bibliographic references in harvard style, live circulation information, links to the catalog record, links to online versions of the item (if applicable), a ready-to-import ris description of the item, and even images of the book cover. a list of classes majax supports is provided in table . examples figure provides an example use of majax widgets. four <span> tags expand into the book cover, a complete harvard-style reference, the valid of a specific marc field ( ), and a display of the current availability of the item, wrapped in a link to the catalog record. texts such as “copy is available” shown in figure are localizable. even though there are multiple majax <span> tags that refer to the same isbn, the majax widget library will contact the majax or majax web service only once per identifier, independent of how often it is used in a page. to manage the load, the majax client site library can be configured to not exceed a maximum number of requests per second, per client. all software described in this paper is available under the lgpl open source license. the majax libraries have been used by us and others for about two years. for instance, the “new books” list in our library uses majax to provide circulation information. faculty members at our institution are using majax to enrich their course websites. a number of libraries have adopted majax , which is particularly easy to host because no additional server is required. ■■ related work most ilss in use today do not provide suitable web-services interfaces to access either bibliographic information figure . architecture of the majax web service information technology and libraries | june or availability data. this shortcoming is addressed by multiple initiatives. the ils discovery interface task force (ils-di) created a set of rec- ommendations that facilitate the integration of discovery interfaces with legacy ilss, but does not define a concrete api. related, the iso holdings standard describes an xml schema to describe the availability of items across sys- tems, but does not describe an api for accessing them. many ilss provide a z . interface in addition to their html- based web opacs, but z . does not provide standardized holdings and availability. nevertheless, there is hope within the community that ils vendors will react to their customers’ needs and provide web-services interfaces that implement these recommenda- tions. the jangle project provides an api and an implementation of the ils-di recommendations through a representations state transfer (rest)–based interface that uses the atom publishing protocol (app). jangle can be linked to legacy ilss via connec- tors. the use of the xml-based app prevents direct access from client-side javascript code, how- ever. in the future, adoption and widespread implementation of the w c working draft on cross- origin resource sharing may relax the same-origin restriction in a controlled fashion, and thus allow access to app feeds from javascript across domains. screen-scraping is a common technique used to over- come the lack of web-services interfaces. for instance, oclc’s worldcat local product obtains access to avail- ability information from legacy ilss in a similar fashion as our majax service. whereas the web services used or created in our work exclusively use a rest-based model and return data in json format, interfaces based on soap (formerly simple object access protocol) whose semantics are described by a wsdl specification provide an alternative if access from within client-side javascript code is not required. html written by adapter <table width=“ ”><tr><td> <span class=“majax-syndetics-vtech” title=“i ”></span> </td><td> <span class=“majax-harvard-reference” title=“i ”></span> <br /> isbn: <span class=“majax-marc- ” title=“i ”></span> <br /> <span class=“majax-linktocatalogmajax-showholdings” title=“i ”></span> </td></tr></table> display in browser after processing dahl, mark., banerjee, kyle., spalti, michael., , digital libraries : integrating content and systems / oxford, chandos publishing, xviii, p. isbn: (hbk.) copy is available figure . example use of majax widgets oclc grid services provides rest-based web-services interfaces to several databases, including the worldcat search api and identifier services such as xisbn, xissn, and xoclcnum for frbr-related metadata. these ser- vices support xml and json and could benefit from widgetization for easier inclusion into client pages. the use of html markup to encode processing instructions is common in javascript frameworks, such as yui or dojo, which use <div> elements with custom- defined attributes (so-called expando attributes) for this purpose. google gadgets uses a similar technique as well. the widely used context objects in spans (coins) specification exploits <span> tags to encode openurl table . selected majax classes majax class replacement majax-marc-fff-s majax-marc-fff majax-syndetics-* majax-showholdings majax-showholdings-brief majax-endnote majax-ebook majax-linktocatalog majax-harvard-reference majax-newline majax-space marc field fff, subfields concatenation of all subfields in field fff book cover image current holdings and availability information …in brief format ris version of record link to online version, if any link to record in catalog reference in harvard style newline space web services and widgets for library information systems | back and bailey techniques for the seamless inclusion of information from web services into websites. we considered the cases where an opac is either the target of such integra- tion or the source of the information being integrated. we focused on client-side techniques in which each user’s browser contacts web services directly because this approach lends itself to the creation of html widgets. these widgets allow the integration and customization of web services without requiring programming. therefore nonprogrammers can become mash-up creators. we described in detail the functionality and use of several widget libraries and web services we built. table provides a summary of the functionality and hosting requirements for each system discussed. although the specific requirements for each system differ because of their respective nature, all systems are designed to be deployable with minimum effort and resource require- ments. this low entry cost, combined with the provision of a high-level, nonprogramming interface, constitute two crucial preconditions for the broad adoption of mash-up techniques in libraries, which in turn has the potential to context objects in pages for processing by client-side extension. librarything uses client-side mash-up tech- niques to incorporate a social tagging service into opac pages. although their technique uses a <div> ele- ment as a placeholder, it does not allow customization via classes—the changes to the content are encoded in custom-generated javascript code for each library that subscribes to the service. the juice project shares our goal of simplifying the enrichment of opac pages with content from other sources. it provides a set of reusable components that is directed at javascript programmers, not librarians. in the computer-science community, multiple emerg- ing projects investigate how to simplify the creation of server-side data mash-ups by end user programmers. ■■ conclusion this paper explored the design space of mash-up table . summary of features and requirements for the widget libraries presented in this paper majax majax google book classes tictoclookup classes web service screen scraping iii record display json proxy for iii record display google book search dynamic link api books.google.com tictoc cloud application tictoclookup .appspot.com hosted by existing millennium installation /screens wsgi/python script on libx.lib.vt.edu google, inc. google, inc. via google app engine data provenance your opac your opac google jisc (www.tictocs .ac.uk) additional cost n/a can use libx.lib.vt.edu for testing, must run wsgi-enabled web server in production free, but subject to google terms of service generous free quota, pay per use beyond that same domain restriction yes no no no widgetization majax.js: class-based: majax- classes gbsclasses.js:class- based: gbs- tictoc.js:class-based: tictoc- requires javascript programming no no no no requires additional server no yes (apache+mod_wsgi) no no (if using gae), else need apache+mod_wsgi iii bibrecord display n/a n/a yes yes iii webbridge integration yes yes yes yes information technology and libraries | june vastly increase the reach and visibility of their electronic resources in the wider community. references . nicole engard, ed., library mashups—exploring new ways to deliver library data (medford, n.j.: information today, ); andrew darby and ron gilmour, “adding delicious data to your library website,” information technology & libraries , no. ( ): – . . monica brown-sica, “playing tag in the dark: diagnosing slowness in library response time,” information technologies & libraries , no. ( ): – . . dapper, “dapper dynamic ads,” http://www.dapper .net/ (accessed june , ); yahoo!, “pipes,” http://pipes .yahoo.com/pipes/ (accessed june , ). . jennifer bowen, “metadata to support next-genera- tion library resource discovery: lessons from the extensible catalog, phase ,” information technology & libraries , no. ( ): – ; john blyberg, “ils customer bill-of-rights,” online posting, blyberg.net, nov. , , http://www.blyberg .net/ / / /ils-customer-bill-of-rights/ (accessed june , ). . douglas crockford, “the application/json media type for javascript object notation (json),” memo, the inter- net society, july , http://www.ietf.org/rfc/rfc .txt (accessed mar. , ). . google, “who’s using the book search apis?” http:// code.google.com/apis/books/casestudies/ (accessed june , ). . innovative interfaces, “millennium ils,” http://www.iii .com/products/millennium_ils.shtml (accessed june , ). . joint information systems committee, “tictocs jour- nal tables of contents service,” http://www.tictocs.ac.uk/ (accessed june , ). . mark dahl, kyle banarjee, and michael spalti, digital libraries: integrating content and systems (oxford, united king- dom: chandos, ). . john ockerbloom et al., “dlf ils discovery interface task group (ils-di) technical recommendation,” (dec. , ), http://diglib.org/architectures/ilsdi/dlf_ils_ discovery_ . .pdf (accessed june , ). . international organization for standardization, “information and documentation—schema for holdings information,” http://www.iso.org/iso/catalogue_detail .htm?csnumber= (accessed june , ) . national information standards organization, “ansi/ niso z . —information retrieval: application service defi- nition and protocol specification,” (bethesda, md.: niso pr., ), http://www.loc.gov/z /agency/z - - .pdf (accessed may , ). . ross singer and james farrugia, “unveiling jangle: untangling library resources and exposing them through the atom publishing protocol,” the code lib journal no. (sept. , ), http://journal.code lib.org/articles/ (accessed apr. , ); roy fielding, “architectural styles and the design of network-based software architectures” (phd diss., university of california, irvine, ); j. c. gregorio, ed., “the atom pub- lishing protocol,” memo, the internet engineering task force, oct. , http://bitworking.org/projects/atom/rfc .html (accessed june , ). . world wide web consortium, “cross-origin resource sharing: w c working draft march ,” http://www .w .org/tr/access-control/ (accessed june , ). . oclc online computer library center, “worldcat and cataloging documentation,” http://www.oclc.org/support/ documentation/worldcat/default.htm (accessed june , ). . f. curbera et al., “unraveling the web services web: an introduction to soap, wsdl, and uddi,” ieee internet comput- ing , no. ( ): – . . oclc online computer library center, “oclc web services,” http://www.worldcat.org/devnet/wiki/services (accessed june , ); international federation of library asso- ciations and institutions study group on the functional require- ments for bibliographic records, “functional requirements for bibliographic records : final report,” http://www.ifla.org/files/ cataloguing/frbr/frbr_ .pdf (accessed mar. , ). . yahoo!, “the yahoo! user interface library (yui),” http://developer.yahoo.com/yui/ (accessed june , ); dojo foundation, “dojo—the javascript toolkit,” http://www .dojotoolkit.org/ (accessed june , ). . google, “gadgets.* api developer’s guide,” http://code. google.com/apis/gadgets/docs/dev_guide.html (accessed june , ). . daniel chudnov, “coins for the link trail,” library jour- nal ( ): – . . librarything, “librarything,” http://www.librarything .com/widget.php (accessed june , ). . robert wallis, “juice—javascript user interface compo- nentised extensions,” http://code.google.com/p/juice-project/ (accessed june , ). . jeffrey wong and jason hong, “making mashups with marmite: towards end-user programming for the web” confer- ence on human factors in computing systems, san jose, california, april –may , : conference proceedings, volume (new york: association for computing machinery, ): – ; guiling wang, shaohua yang, and yanbo han, “mashroom: end-user mashup programming using nested tables” (paper presented at the international world wide web conference, madrid, spain, ): – ; nan zang, “mashups for the web-active user” (paper presented at the ieee symposium on visual languages and human-centric computing, herrshing am ammersee, germany, ): – . tending a wild garden: library web design for persons with disabilities | vandenbark r. todd vandenbark tending a wild garden: library web design for persons with disabilities nearly one-fifth of americans have some form of dis- ability, and accessibility guidelines and standards that apply to libraries are complicated, unclear, and difficult to achieve. understanding how persons with disabilities access web-based content is critical to accessible design. recent research supports the use of a database-driven model for library web development. existing tech- nologies offer a variety of tools to meet disabled patrons’ needs, and resources exist to assist library professionals in obtaining and evaluating product accessibility infor- mation from vendors. librarians in charge of technology can best serve these patrons by proactively updating and adapting services as assistive technologies improve. i n march , eighty-two countries signed the united nations’ convention on the rights of persons with disabilities, including canada, the european community, and the united states. the convention’s purpose was “to promote, protect and ensure the full and equal enjoyment of all human rights and fundamental freedoms by all persons with disabilities, and to promote respect for their inherent dignity.” among the many proscriptions for assuring respect and equal treatment of people with disabilities (pwd) under the law, signatories agreed to take appropriate measures: (g) to promote access for persons with disabilities to new information and communications technolo- gies and systems, including the internet; and (h) to promote the design, development, production and distribution of accessible information and communications technologies and systems at an early stage, so that these technologies and systems become accessible at minimum cost. in addition, the convention seeks to guarantee equal access to information by doing the following: (c) urging private entities that provide services to the general public, including through the internet, to provide information and services in accessible and usable formats for persons with disabilities; and (d) encouraging the mass media, including providers of information through the internet, to make their services accessible to persons with disabilities. because the internet and its design standards are evolv- ing at a dizzying rate, it is difficult to create websites that are both cutting-edge and standards-compliant. this paper evaluates the challenge of web design as it relates to individuals with disabilities, exploring current standards, and offering recommendations for accessible development. examining the provision of it for this demographic is vital because according to the u.s. census bureau, the u.s. public includes about . mil- lion noninstitutionalized people living with disabilities, . million of which are severely disabled. this means that nearly one-fifth of the u.s. public faces some physi- cal, mental, sensory, or other functional impairment ( percent in ). because a library’s mandate is to make its resources accessible to everyone, it is important to attend to the special challenges faced by patrons with disabilities and to offer appropriate services with those special needs in mind. n current u.s. regulations, standards, and guidelines in congress enacted the americans with disabilities act (ada), the first comprehensive legislation mandating equal treatment under the law for pwd. the ada pro- hibits discrimination against pwd in employment, public services, public accommodations, and in telecommunica- tions. title ii of the ada mandates that all state govern- ments, local governments, and public agencies provide access for pwd to all of their activities, services, and programs. since school, public, and academic libraries are under the purview of title ii, they must “furnish auxiliary aids and services when necessary to ensure effective com- munication.” though predating widespread use of the internet, the law’s intent points toward the adoption and adaptation of appropriate technologies to allow persons with a variety of disabilities to access electronic resources in a way that is most effective for them. changes to section of the rehabilitation act enacted in and introduced the first standards for “accessible information technology recognized by the federal government.” many state and local govern- ments have since passed laws applying the standards of section to government agencies and related services. according to the access board, the independent federal agency charged with assuring compliance with a variety of laws regarding services to pwd, information and com- munication technology (ict) includes any equipment or interconnected system or subsystem of equipment, that is used in the creation, conversion, or duplication of data or information. the term electronic r. todd vandenbark (todd.vandenbark@utah.edu) is web ser- vices librarian, eccles health sciences library, university of utah, salt lake city. information technology and libraries | march and information technology includes, but is not limited to, telecommunications products (such as telephones), information kiosks and transaction machines, world wide web sites, multimedia, and office equipment such as copiers and fax machines. the access board further specifies guidelines for “web-based intranet and internet information and appli- cations,” which are directly relevant to the provision of such services in libraries. what follows is a detailed examination of these standards with examples to assist in understanding and implementation. (a) a text equivalent for every non-text element shall be provided. assistive technology cannot yet describe what pictures and other images look like; they require meaningful text-based information asso- ciated with each picture. if an image directs the user to do something, the associated text must explain the purpose and meaning of the image. this way, someone who cannot see the screen can understand and navigate the page success- fully. this is generally accomplished by using the “alt” and “longdesc” attributes for images: <img src=“image.jpg” alt=“short description of image.” longdesc=“explanation.txt” />. however, these aids also can clutter a page when not used properly. the current versions of the most popular screen-reader software do not limit the amount of “alt” text they can read. however, freedom scientific’s jaws .x divides the “alt” attribute into distinct chunks of characters each (excluding spaces) and reads them separately as if they were separate graphics. this can be confusing to the end user. longer con- tent can be put into a separate text file and the file linked to using the “longdesc” attribute. when a page contains audio or video files, a text alternative needs to be provided. for audio files such as inter- views, lectures, and podcasts, a link to a transcript of the audio file must be immediately available. for video clips such as those on youtube, captions must accompany the clip. (b) equivalent alternatives for any multimedia presen- tation shall be synchronized with the presentation. this means that captions for video must be real-time and synchronized with the actions in the video, not contained solely in a separate transcript. (c) web pages shall be designed so that all informa- tion conveyed with color is also available with- out color, for example from context or markup. while color can be used, it cannot be the sole source or indicator of information. imagine an edu- cational website offering a story problem presented in black and green print, and the answer to the problem could be deciphered using only the green letters. this would be inaccessible to students who have certain forms of color-blindness as well as those who use screen-reader software. (d) documents shall be organized so they are read- able without requiring an associated style sheet. the introduction of cascading style sheets (css) can improve accessibility because they allow the separation of presentation from content. however, not all browsers fully support css, so webpages need to be designed so any browser can read them accurately. the content needs to be organized so that it can be read and understood with css for- matting turned off. (e) redundant text links shall be provided for each active region of a server-side image map, and (f) client-side image maps shall be provided instead of server-side image maps except where the regions cannot be defined with an available geometric shape. an image map can be thought of as a geometri- cally defined and arranged group of links to other content on a site. a clickable map of the fifty u.s. states is an example of a functioning image map. a server-side image map would appear to a screen reader only as a set of coordinates, whereas client- side maps can include information about where the link leads through “alt” text. the best practice is to only use client-side image maps and make sure the “alt” text is descriptive and meaningful. (g) row and column headers shall be identified for data tables, and (h) markup shall be used to associate data cells and header cells for data tables that have two or more logical levels of row or column headers. correct table coding is critical. each table should use the “table summary” attribute to provide a meaningful description of its content and arrange- ment: <table summary=“concise explanation belongs here.”>. headers should be coded using the table header (“th”) tag, and its “scope” attri- bute should specify whether the header applies to a row or a column: <th scope=“col”> or <th scope=“row”>. if the table’s content is complex, it may be necessary to provide an alternative presen- tation of the information. it is best to rely on css for page layout, taking into consideration the direc- tions in subparagraph (d) above. (i) frames shall be titled with text that facili- tates frame identification and navigation. frames are a deprecated feature of html, and their use should be avoided in favor of css layout. (j) pages shall be designed to avoid caus- ing the screen to flicker with a frequency greater than hz and lower than hz. lights with flicker rates in this range can trigger epileptic seizures. blinking or flashing elements on tending a wild garden: library web design for persons with disabilities | vandenbark a webpage should be avoided until browsers pro- vide the user with the ability to control flickering. (k) a text-only page, with equivalent information or functionality, shall be provided to make a web site comply with the provisions of this part, when compliance cannot be accomplished any other way. the content of the text-only page shall be updated whenever the primary page changes. complex content that is entirely visual in nature may require a separate text-only page, such as a page showing the english alphabet in american sign language. this requirement also serves as a stopgap measure for existing sites that require reworking for accessibility. some consider this to be the web’s version of separate-but-equal ser- vices, and should be avoided. offering a text-only alternative site can increase the sense of exclusion that pwd already feel. also, such versions of a website tend not to be equivalent to the parent site, leaving out promotions or advertisements. finally, a text-only version increases the workload of web development staff, making them more costly than creating a single, fully accessible site in the first place. (l) when pages utilize scripting languages to display content, or to create interface elements, the informa- tion provided by the script shall be identified with functional text that can be read by assistive technology. scripting languages such as javascript allow for more interactive content on a page while reducing the number of times the computer screen needs to be refreshed. if functional text is not available, the screen reader attempts to read the script’s code, which outputs as a meaningless jumble of charac- ters. using redundant text links avoids this result. (m) when a web page requires that an applet, plug-in, or other application be present on the client system to interpret page content, the page must provide a link to a plug-in or applet that complies with [subpart b: technical standards] § . (a) through (i). web developers need to ascertain whether a given plug-in or applet is accessible before requiring their webpage’s visitors to use it. when using applications such as quicktime or realaudio, it is important to provide an accessible link on the same page that will allow users to install the necessary plug-in. (n) when electronic forms are designed to be completed on-line, the form shall allow people using assistive technology to access information, field elements, and functionality required for completion and submis- sion of the form, including all directions and cues. if scripts used in the completion of the form are inaccessible, an alternative method of completing the form must be made immediately available. each element of a form needs to be labeled prop- erly using the <label> tag. (o) a method shall be provided that per- mits users to skip repetitive navigation links. persons using screen reader software typically navigate through pages using the tab key, listen- ing as the text is read aloud. websites commonly place their logo at the top of each page and make this graphic a link to the site’s homepage. many sites also use a line of graphic images just beneath this logo on every page to serve as a navigation bar. to avoid having to listen through this same list of links on every page just to get to the page’s content, a “skip to content” link as the first option at the top of each page provides a simple solution to this problem. (p) when a timed response is required, the user shall be alerted and given sufficient time to indicate more time is required. some sites log a user off if they have not typed or otherwise interacted with the page after a certain time period. users must be notified in advance that this is going to happen and given sufficient time to respond and request more time as needed. n standards-setting groups and their work one organization that seeks to move internet tech- nology beyond basic section compliance is the web accessibility initiative (wai) of the world wide web consortium (w c). the mission of the wai is to develop n guidelines that are widely regarded as the interna- tional standard for web accessibility; n support materials to help understand and imple- ment web accessibility; and n resources through international collaboration. the w c published its first web content accessibility guidelines (wcag . ) in may of for making online content accessible to pwd. by following these guidelines, developers create web content that is readily available to every user regardless of the way it’s accessed. the wai provides ten quick tips for improving accessibility in website design: n images and animations. use the “alt” attribute to describe the function of each visual. n image maps. use the client-side map and text for hotspots. n multimedia. provide captioning and transcripts of audio, and descriptions of video. information technology and libraries | march n hypertext links. use text that makes sense when read out of context. for example, avoid “click here.” n page organization. use headings, lists, and consis- tent structure. use css for layout and style where possible. n graphs and charts. summarize or use the “longdesc” attribute. n scripts, applets, and plug-ins. provide alternative content in case active features are inaccessible or unsupported. n frames. use the “noframes” element and meaning- ful titles. n tables. make line-by-line reading sensible. summarize. n check your work. validate. use tools, checklist, and guidelines at http://www.w .org/tr/wcag. many libraries and other organizations have sought to follow wcag . since it was published. recently, the w c updated their standards to wcag . , and the wai website offers an overview of these guidelines along with a “customizable quick reference” designed to facilitate successful compliance. the principles behind . can be summarized by the acronym p.o.u.r. perceivable n provide text alternatives for non-text content. n provide captions and alternatives for multimedia. n make information adaptable and available to assistive technologies. n use sufficient contrast to make things easy to see and hear. operable n make all functionality keyboard accessible. n give users enough time to read and use content. n do not use content known to cause seizures. n help users navigate and find content. understandable n make text readable and understandable. n make content appear and operate in predictable ways. n help users avoid and correct mistakes. robust n maximize compatibility with current and future technologies. these guidelines offer assistance in creating acces- sible web-based materials. given their breadth, however, they raise concerns of overly wide interpretation and the strong possibility of falling short of section standards. reading the details in wcag . does not give any additional assistance to library web developers on how to create a section –compliant website. clark points out that the three wcag . documents are long ( – pages), confusing, and sometimes internally contradic- tory. the goal of a library webmaster is to provide an interface (website, opac, database, and so on) that is both cutting-edge and accessible, and to encourage its use by patrons of all ability levels. while they have outlined a helpful rationale, the w c’s overlong guidelines do little to help library web developers to achieve this goal. n recommendations libraries today typically offer three types of web-based resources: ( ) access to the internet, ( ) access to subscrip- tion databases, and ( ) a library’s own webpage, all of which need to be accessible to pwd. libraries trying to comply with section are required to “furnish auxil- iary aids and services when necessary to ensure effective communication.” there are a number of options avail- able to libraries on tight budgets. the first set involves the features built into each computer’s operating sys- tem and software. for some users with visual impair- ments, enlarging the font size of text and images on the screen will make electronic content more accessible. both macintosh and windows system software have universal-access capabilities built in, including the ability to read aloud text that is on the screen using synthesized speech. the mac read-aloud tool is called voice over; the windows read-aloud tool is called narrator. both systems allow for screen magnification. exploring and learning the capabilities of these systems to enhance accessibility is a free and easy first step for any library’s technology offerings, regardless of funding restrictions. libraries with more substantial technology budgets have a wide variety of hardware and software options to choose from to meet the needs of pwd. for patrons with visual impairments, several software packages are available to read aloud the content of a website or other electronic document using synthesized speech. jaws by freedom scientific and windoweyes by gw micro are two of the best-known software packages, and both include the ability to output to a refreshable braille dis- play (which both companies also sell). kurzweil is an education-oriented software package that not only reads on-screen text aloud but has a wealth of additional tools to assist students with learning difficulties such as attention deficit disorder or dyslexia. it is designed to integrate with any education package as well as to assist students whose primary language is not english. persons with low vision needing screen magnification beyond the features windows offers may look to magic by freedom scientific or zoomtext by ai squared. some of these tending a wild garden: library web design for persons with disabilities | vandenbark software companies offer free trial versions, have online demonstrations, or both. because prices for this software and related equipment can be high, it is prudent to first check with patrons with visual impairments and profes- sionals in the field prior to making your purchase. humbert and stores, members of indiana university’s web accessibility team, offer accessibility evaluations of websites and other services at the university. when asked to compare windows and macintosh systems as to their usefulness in assisting pwd with web-based media, humbert rated the windows operating system superior, explaining that it has the proper “handles” coded into its software for screen readers and assistive technologies to grab onto. assistive technology software is more stable in windows vista because its predecessor, windows xp, “used hacked together drivers to display the informa- tion.” humbert discourages the use of vista and jaws on an older machine because vista is a memory hog and can crash jaws along with the rest of the system. the web browsers internet explorer and firefox allow the user to enlarge text and images on a webpage, though firefox is more effective. text can be enlarged only if the webpage being viewed is designed using resizable fonts. stores, who is profoundly visually impaired, uses jaws screen-reader software to work and to surf the web. she notes that both browsers work equally well with screen- reader software. an important web-based resource that libraries pro- vide is subscription databases. however, as one study has shown, “most librarians lack the time, resources and/or skills to evaluate the degree to which their library subscription databases are accessible to their disability communities.” the question is do the vendors them- selves make an effort to produce an accessible product? a survey of twelve major database companies found that while most “have integrated accessibility standards/ guidelines into their search interfaces and/or plan to improve accessibility in future releases,” only five actu- ally conducted usability studies with people who use assistive technology. a number of studies have found that “while most databases are functionally accessible, com- panies need to do more to meet the needs of the disability community and assure librarians of the accessibility of their products.” subscription databases can be inaccessible to pwd in the display of search results and accompanying infor- mation. the three most common forms of results deliv- ery are html full text, html full text with graphics, and pdf files. pdf files are notoriously inaccessible to persons using screen readers. while adobe has made significant strides in rendering pdfs accessible, many databases contain numerous pdf documents created in versions of adobe acrobat prior to version . (released in ), which are not properly tagged for screen read- ers. even newer pdf documents are only as accessible as their tagging allows. journal articles received from publishers may or may not be properly tagged, so data- base companies cannot guarantee that their content is fully accessible. one vendor that is avoiding this trap is jstor. using optical character recognition (ocr) soft- ware, jstor delivers image-based pdfs with embedded text to make their content available to screen readers. librarians must insist that database packages be acces- sible and compatible with the forms of assistive technol- ogy most frequently used by their patrons, both in-house and online. one tool used to evaluate database (or other prod- uct) accessibility is the voluntary product accessibility template (vpat). created in partnership between the information technology industry (iti) council and the u.s. general services administration (gsa) in , it provides “a simple, internet-based tool to assist federal contracting and procurement officials in fulfilling the new market research requirements contained in the section implementing regulations.” vpat is a voluntary disclosure form arranged in a series of tables listing the criteria of relevant subsections of section discussed previously. blank cells are provided to allow company representatives to describe how their product’s support- ing features meet the criteria and to provide additional detailed information. library personnel can request that vendors complete this form to document which sub- sections of section their products meet, and how. to be most useful, the form needs to be completed by company representatives with both a clear understand- ing of section and its technical details and thorough knowledge of their product. knowledgeable library staff are encouraged to verify the quality and accuracy of the information provided before purchasing. like databases, a library’s website needs to be acces- sible to patrons with a variety of needs. according to muncaster, accessible sites are percent easier for every- one to use and are more likely to be found by internet search engines. fully accessible websites are simpler to maintain and are on average percent smaller than inaccessible ones, which means they download faster, making them easier to use. in creating a basic site, cur- rent best practice has been to render the content in html or xhtml and design the layout using css. this way, if it is discovered the site’s pages are not fully accessible, a simple change to the css updates all pages, saving the site manager time and effort. finally, creating an acces- sible site from the beginning is substantially easier than retrofitting an old one. a complete rebuild of a library website is an opportu- nity to improve accessibility. reynolds’ article on creating a user-centered website for the johnson county (kans.) library offers an example of how libraries can apply basic information architecture design principles on a budget. johnson county focused on simple, low-budget information technology and libraries | march usability studies involving patrons in the selection of site navigation categories, designing the layout, and testing the resulting user interface. by involving average users in this process, this library was able to achieve substantial improvements in the site’s usability. prior to the redesign, usability testing determined that percent of users were not successful in finding information on the library’s old site. after the redesign, “only % of patrons were unsuccessful in finding core-task information on the first attempt.” even so, a quick test of the site with the online accessibility evaluation tool cynthiasays indicates that it still does not fully meet the requirements of section . had the library’s staff included pwd in their process, the demonstrated degree of improvement might have allowed them to meet and possibly exceed this standard. an understanding of how a person with disabilities experiences the online environment can help point the way toward improved accessibility. a recent study in the united kingdom tracked the eye movements of able- bodied computer users in an effort to answer these ques- tions. researchers asked eighteen people with normal or corrected vision to search for answers on two versions of a bbc website—the standard graphical page and the text- only version. subjects’ eyes tended to dart around the standard page “as they attempt to locate what appears visually to be the next most likely location” for the answer. but in searching the text-only page, subjects went line-by-line, making smaller jumps across each page. researchers determined that the webpage and its layout serve as a form of external memory, providing visual cues to the structure of its content and how to navigate it. if the internet is an information superhighway, then the layout of a standard webpage serves as the borders and directional signs for browsing. the visual cues and navigation aids inherent in cur- rent webpages’ layouts provide no auditory equivalent for presentation to people with visual impairments. information seeking on the web is a complex process requiring “the ability to switch and coordinate among multiple information-seeking strategies” such as brows- ing, scanning, query-based searching, and so on. if web browsers could translate formatting and presentation into audio tailored to the needs of the visually impaired, the use of the internet would be a far more satisfying experi- ence for those users. however, such web programming would require years of additional research and develop- ment. in the meantime, web librarians must strive to build sites that are clean, hierarchical, and usable by all persons by following to the standards and guidelines currently available. one way to enhance the accessibility of sites is to fol- low a database-driven web development model. in addi- tion to using xhtml and css, dunlap recommends that content be stored in a relational database such as mysql and that a coding language such as php be used to create pages dynamically. this approach has two advantages. first, it allows for the creation of “a flexible website design style that lives in a single, easily modified file that controls the presentation of every web page of the site.” second, it requires far less time for site maintenance, freeing staff to devote time to assuring accessibility while accommodating changes in web technology. such a model can be used by database vendors to ensure that their services can seamlessly integrate with the library’s online content. use of mobile phones and similar devices to browse the web is at an all-time high, and content providers are eager to make their sites mobile-friendly. many of these end users experience similar barriers to accessing this content as pwd do. for example, persons with some motor disabilities as well as mobile phones with only a numeric keypad cannot access sites with links requiring the use of a mouse. sites that follow either the w c’s mobile web best practices (mwbp) or wcag are well on their way to meeting both standards. by properly asso- ciating labels with their controls, internet content can be made fully accessible to both end users. understanding the similarities between mwpb and wcag can lead to website design that is truly perceivable, operable, under- standable, and robust. n summary librarians with responsibility for web design and tech- nology management operate in an evolving environment. legal requirements make clear the expectation to serve the wide variety of needs of patrons with disabilities. yet the guidelines and standards available to assist in this venture range from complex to vague and insufficient. assistive technologies continue to improve with many traditional vendors confident that their products are accessible. in actual use, however, substantial challenges and shortcomings remain. the challenge for technology librarians is to be proactive in keeping abreast of tech- nological advances, to experiment and learn from their efforts, and to continually update and adapt to provide web or hypermedia information and services to patrons of all kinds. references . united nations, convention on the rights of persons with disabilities, , http://www.un.org/disabilities/default .asp?navid= &pid= (accessed aug. , ). . ibid. . erika steinmetz, americans with disabilities (washington, d.c.: u.s. census bureau, ). tending a wild garden: library web design for persons with disabilities | vandenbark . u.s. department of justice, civil rights division, disabil- ity rights section, “title ii highlights,” aug. , , http:// www.ada.gov/t hlt .htm (accessed july , ). . marilyn irwin, resources and services for people with dis- abilities: lesson b transcript (indianapolis: indiana university at indianapolis school of library and information science, ): . ibid., . amendment to section of the rehabilitation act, subpart b—technical standards, § . , http://www .section .gov/index.cfm?fuseaction=content&id= #appli cation (access dec. , ). . access it, “how long can an ‘alt’ attribute be?” uni- versity of washington, , http://www.washington.edu/ accessit/articles? (accessed dec. , ). . matt may, “on ‘separate but equal’ design,” online post- ing, june , , bestkungfu weblog, http://www.bestkungfu .com/archive/date/ / /on-separate-but-equal-design/ (accessed dec. , ). . web accessibility initiative, “wai mission and organiza- tion,” , http://www.w .org/wai/about.html (accessed july , ). . shawn lawton henry and pasquale popolizio, “wai, quick tips to make accessible web sites,” world wide web consortium, feb. , , http://www.w .org/wai/quicktips/ overview.php (accessed mar. , ). . ben caldwell et al., “web content accessibility guide- lines (wcag) . ,” world wide web consortium, dec. , , http://www.w .org/tr/wcag / (accessed july , ). . joe clark, “to hell with wcag ,” a list apart no. (may , ), http://www.alistapart.com/articles/tohellwith wcag (accessed july , ). . u.s. department of justice, “title ii highlights.” . joseph a. humbert and mary stores, questions about new software and accessibility (richmond, ind., july , ). . ibid. . s. l. byerley, m. b. chambers, and m. thohira, “acces- sibility of web-based library databases: the vendors’ perspec- tives in ,” library hi tech , no. ( ): – . . ibid. . p. muncaster, “poor accessibility has a price,” vnu net, feb. , , http://www.vnunet.com/articles/send/ (accessed july , ). . information technology industry council, “faq: volun- tary product accessibility template (vpat),” http://www.itic .org/archives/articles/ /faq_voluntary_product_ accessibility_template_vpat.php (accessed july , ). . muncaster, “poor accessibility has a price.” . isaac hunter dunlap, “how database-driven web sites enhance accessibility,” library hi tech , no. ( ): – . . erica reynolds, “the secret to patron-centered web design: cheap, easy, and powerful usability techniques,” com- puters in libraries , no. ( ): – . . caroline jay et al., “how people use presentation to search for a link: expanding the understanding of accessibility on the web,” universal access in the information society , no. ( ): – . . c. kouroupetroglou, m. salampasis, and a. manitsaris, “browsing shortcuts as a means to improve information seek- ing of blind people in the www,” universal access in the informa- tion society , no. ( ): . . dunlap, “how database-driven web sites enhance accessibility.” . web accessibility initiative, “mobile web best prac- tices . ,” july , , http://www.w .org/tr/mobile-bp (accessed aug. , ). information technology and libraries | march paul t. jaeger and zheng yan one law with two outcomes: comparing the implementation of cipa in public libraries and schools though the children’s internet protection act (cipa) established requirements for both public libraries and public schools to adopt filters on all of their computers when they receive certain federal funding, it has not attracted a great amount of research into the effects on libraries and schools and the users of these social insti- tutions. this paper explores the implications of cipa in terms of its effects on public libraries and public schools, individually and in tandem. drawing from both library and education research, the paper examines the legal background and basis of cipa, the current state of internet access and levels of filtering in public librar- ies and public schools, the perceived value of cipa, the perceived consequences of cipa, the differences in levels of implementation of cipa in public libraries and public schools, and the reasons for those dramatic differences. after an analysis of these issues within the greater policy context, the paper suggests research questions to help provide more data about the challenges and questions revealed in this analysis. t he children’s internet protection act (cipa) estab- lished requirements for both public libraries and public schools to—as a condition for receiving cer- tain federal funds—adopt filters on all of their computers to protect children from online content that was deemed potentially harmful. passed in , cipa was initially implemented by public schools after its passage, but it was not widely implemented in public libraries until the supreme court decision (united states v. american library association) upholding the law’s constitutional- ity. now that cipa has been extensively implemented for five years in libraries and eight years in schools, it has had time to have significant effects on access to online information and services. while the goal of filter- ing requirements is to protect children from potentially inappropriate content, filtering also creates major edu- cational and social implications because filters also limit access to other kinds of information and create different perceptions about schools and libraries as social institu- tions. curiously, cipa and its requirements have not attracted a great amount of research into the effects on schools, libraries, and the users of these social institu- tions. much of the literature about cipa has focused on practical issues—either recommendations on implement- ing filters or stories of practical experiences with filtering. while those types of writing are valuable to practitioners who must deal with the consequences of filtering, there are major educational and societal issues raised by filter- ing that merit much greater exploration. while relatively small bodies of research have been generated about cipa’s effects in public libraries and public schools, thus far these two strands of research have remained separate. but it is the contention of this paper that these two strands of research, when viewed together, have much more value for creating a broader understanding of the educational and societal implications. it would be impossible to see the real consequences of cipa without the development of an integrative picture of its effects on both public schools and public libraries. in this paper, the implications of cipa will be explored in terms of effects on public libraries and public schools, individually and in tandem. public libraries and public schools are generally considered separate but related public sphere entities because both serve core educa- tional and information-provision functions in society. furthermore, the fact that public schools also contain school library media centers highlights some very inter- esting points of intersection between public libraries and school libraries in terms of the consequences of cipa: while cipa requires filtering of computers throughout public libraries and public schools, the presence of school library media centers makes the connection between libraries and schools stronger, as do the teaching roles of public libraries (e.g., training classes, workshops, and evening classes). n the legal road to cipa history under cipa, public libraries and public schools receiving certain kinds of federal funds are required to use filtering programs to protect children under the age of seventeen from harmful visual depictions on the internet and to provide public notices and hearings to increase public awareness of internet safety. senator john mccain (r-az) sponsored cipa, and it was signed into law by president bill clinton on december , . cipa requires that filters at public libraries and public schools block three specific types of content: ( ) obscene material (that paul t. jaeger (pjaeger@umd.edu) is assistant professor at the college of information studies and director of the center for information policy and electronic government of the university of maryland in college park. zheng yan (zyan@uamail.albany .edu) is associate professor at the department of educational and counseling psychology in the school of education of the state university of new york at albany. one law with two outcomes | jaeger and yan which appeals to prurient interests only and is “offen- sive to community standards”); ( ) child pornography (depictions of sexual conduct and or lewd exhibitionism involving minors); and ( ) material that is harmful to minors (depictions of nudity and sexual activity that lack artistic, literary, or scientific value). cipa focused on “the recipients of internet transmission,” rather than the send- ers, in an attempt to avoid the constitutional issues that undermined the previous attempts to regulate internet content. using congressional authority under the spending clause of article i, section of the u.s. constitution, cipa ties the direct or indirect receipt of certain types of federal funds to the installation of filters on library and school computers. therefore each public library and school that receives the applicable types of federal funding must implement filters on all computers in the library and school buildings, including computers that are exclusively for staff use. libraries and schools had to address these issues very quickly because the federal communications commission (fcc) mandated certifi- cation of compliance with cipa by funding year , which began in summer . cipa requires that filters on computers block three specific types of content, and each of the three cat- egories of materials has a specific legal meaning. the first type—obscene materials—is statutorily defined as depicting sexual conduct that appeals only to prurient interests, is offensive to community standards, and lacks serious literary, artistic, political, or scientific value. historically, obscene speech has been viewed as being bereft of any meaningful ideas or educational, social, or professional value to society. statutes regulating speech as obscene have to do so very carefully and specifically, and speech can only be labeled obscene if the entire work is without merit. if speech has any educational, social, or professional importance, even for embody- ing controversial or unorthodox ideas, it is supposed to receive first amendment protection. the second type of content—child pornography—is statutorily defined as depicting any form of sexual conduct or lewd exhi- bitionism involving minors. both of these types of speech have a long history of being regulated and being considered as having no constitutional protections in the united states. the third type of content that must be filtered— material that is harmful to minors—encompasses a range of otherwise protected forms of speech. cipa defines “harmful to minors” as including any depiction of nudity, sexual activity, or simulated sexual activity that has no serious literary, artistic, political, or scientific value to minors. the material that falls into this third category is constitutionally protected speech that encompasses any depiction of nudity, sexual activity, or simulated sexual activity that has serious literary, artistic, political, or scientific value to adults. along with possibly includ- ing a range of materials related to literature, art, science, and policy, this third category may involve materials on issues vital to personal well-being such as safe sexual practices, sexual identity issues, and even general health care issues such as breast cancer. in addition to the filtering requirements, section also prescribes an internet awareness strategy that public libraries and schools must adopt to address five major internet safety issues related to minors. it requires librar- ies and schools to provide reasonable public notice and to hold at least one public hearing or meeting to address these internet safety issues. requirements for schools and libraries cipa includes sections specifying two major strategies for protecting children online (mainly in sections , , , and ) as well as sections describing vari- ous definitions and procedural issues for implementing the strategies (mainly in sections , , , , , and ). section specifies the primary internet protec- tion strategy—filtering—in public schools. specifically, it amends the elementary and secondary education act of by limiting funding availability for schools under section of the communication act of . through a compliance certification process within a school under supervision by the local educational agency, it requires schools to include the operation of a technology protec- tion measure that protects students against access to visual depictions that are obscene, are child pornography, or are harmful to minors under the age of seventeen. likewise, section specifies the same filtering strategy in public libraries. specifically, it amends section of the museum and library service act of / by limiting funding availability for libraries under sec- tion of the communication act of . through a compliance certification process within a library under supervision by the institute of museum and library services (imls), it requires libraries to include the opera- tion of a technology protection measure that protects stu- dents against access to visual depictions that are obscene, child pornography, or harmful to minors under the age of seventeen. section is a requirement for both libraries and schools to enforce the internet safety policy with the internet safety policy strategy and the filtering technol- ogy strategy as a condition of universal service discounts. specifically, it amends section of the communication act of and requests both schools and libraries to monitor the online activities of minors, operate a tech- nical protection measure, provide reasonable public notice, and hold at least one public hearing or meeting to address the internet safety policy. this is through the information technology and libraries | march certification process regulated by the fcc. section , titled the neighborhood children’s internet protection act (ncipa), amends section of the communication act of and requires schools and libraries to adopt and implement an internet safety policy. it specifies five types of internet safety issues: ( ) access by minors to inappropriate matter on the internet; ( ) safety and security of minors when using e-mail, chat rooms, and other online communications; ( ) unauthor- ized access; ( ) unauthorized disclosure, use, and dis- semination of personal information; and ( ) measures to restrict access to harmful online materials. from the above summary, it is clear that ( ) the two protection strategies of cipa (the internet filtering strat- egy and safety policy strategy) were equally enforced in both public schools and public libraries because they are two of the most important social institutions for children’s internet safety; ( ) the nature of the implementation mechanism is exactly the same, using the same federal funding mechanisms as the sole financial incentive (lim- iting funding availability for schools and libraries under section of the communication act of ) through a compliance certification process to enforce the imple- mentation of cipa; and ( ) the actual implementation procedure differs in libraries and schools, with schools to be certified under the supervision of local educational agencies (such as school districts and state departments of education) and with libraries to be certified within a library under the supervision of the imls. economics of cipa the universal service program (commonly known as e–rate) was established by the telecommunications act of to provide discounts, ranging from to percent, to libraries and schools for telecommunications services, internet services, internal systems, and equip- ment. the program has been very successful, provid- ing approximately $ . billion dollars a year to public schools, public libraries, and public hospitals. the vast majority of e-rate funding—about percent—goes to public schools each year, with roughly percent being awarded to public libraries and the remainder going to hospitals. the emphasis on funding schools results from the large number of public schools and the size- able computing needs of all of these schools. but even percent of the e-rate funding is quite substantial, with public libraries receiving more than $ million between and . schools received about $ billion in the same time period. along with e-rate funds, the library services and technology act (lsta) program adminis- tered by the imls provides money to each state library agency to use on library programs and services in that state, though the amount of these funds is considerably lower than e-rate funds. the american library association (ala) has noted that the e-rate program has been particularly significant in its role of expanding online access to students and to library patrons in both rural and underserved com- munities. in addition to the effect on libraries, e-rate and lsta funds have significantly affected the lives of individuals and communities. these programs have contributed to the increase in the availability of free public internet access in schools and libraries. by , more than percent of public school libraries provided students with internet access. by , . percent of public library branches were connected to the internet, and . percent of public library branches offered pub- lic internet access. however, only a small portion of libraries and schools used filters prior to cipa. since the advent of computers in libraries, librarians typically had used informal monitoring practices for computer users to ensure that nothing age inappropriate or morally offensive was publicly visible. some individual school and library systems, such as in kansas and indiana, even developed formal or informal statewide internet safety strategies and approaches. why were only libraries and schools chosen to protect children’s online safety? while there are many social institutions that could have been the focus of cipa, the law places the requirements specifically on public libraries and public schools. if congress was so interested in protecting children from access to harmful internet content, it seems that the law would be more expansive and focused on the content itself rather than filtering access to the content. however, earlier laws that attempted to regulate access to internet content failed legal challenges specifically because they tried to regulate content. prior to the enactment of cipa, there were a num- ber of other proposed laws aimed at preventing minors from accessing inappropriate internet content. the communications decency act (cda) of prohib- ited the sending or posting of obscene material through the internet to individuals under the age of eighteen. however, the supreme court found the cda to be unconstitutional, stating that the law violated free speech under the first amendment. in , congress passed the child online protection act (copa), which prohibited commercial websites from displaying material deemed harmful to minors and imposed criminal penalties on internet violators. a three-panel judge for the district court for the eastern district of pennsylvania ruled that copa’s focus on “contemporary community standards” violated the first amendment, and the panel subsequently imposed an one law with two outcomes | jaeger and yan injunction on copa’s enforcement. cipa’s force comes from congress’s power under the spending clause; that is, congress can legally attach requirements to funds that it gives out. since cipa is based on economic persuasion—the potential loss of funds for technology—the law can only have an effect on recipients of those funds. while regulating internet access in other venues like coffee shops, internet cafés, bookstores, and even individual homes would provide a more comprehensive shield to limit children’s access to certain online content, these institutions could not be reached under the spending clause. as a result, the burdens of cipa fall squarely on public libraries and public schools. n the current state of filtering when did cipa actually come into effect in libraries and schools? after overcoming a series of legal challenges that were ultimately decided by the supreme court, cipa came into effect in full force in , though percent of public schools were already in compliance with cipa in . when the court upheld the constitutionality of cipa, the legal challenge by public libraries centered on the way the statute was written. the court’s deci- sion states that the wording of the law does not place unconstitutional limitations on free speech in public libraries. to continue receiving federal dollars directly or indirectly through certain federal programs, public libraries and schools were required to install filtering technologies on all computers. while the case decided by the supreme court focused on public libraries, the decision virtually precludes public schools from making the same or related challenges. before that case was decided, however, most schools had already adopted filters to comply with cipa. as a result of cipa, a public library or public school must install technology protection measures, better known as filters, on all of its computers if it receives n e-rate discounts for internet access costs, n e–rate discounts for internal connections costs, n lsta funding for direct internet costs, or n lsta funding for purchasing technology to access the internet. the requirements of cipa extend to public libraries, public schools, and any library institution that receives lsta and e–rate funds as part of a system, including state library agencies and library consortia. as a result of the financial incentives to comply, almost percent of public schools in the united states have implemented the requirements of cipa, and approximately half of public libraries have done so. how many public schools have implemented cipa? according to the latest report by the department of education (see table ), by , percent of public schools had implemented both the internet filtering strategy and safety policy strategy. in fact, in (the first year cipa was in effect), percent of schools had implemented cipa, with percent filtering by . when compared to the percentage of all public schools with internet access from to , internet access became nearly universal in schools between and ( to percent), and one can see that the internet access percentage in was almost the same as the cipa implementation percentage. according to the department of education, the above estimations are based on a survey of , elementary and secondary schools selected from , elementary schools and , secondary and combined schools. after reviewing the design and administration of the sur- vey, it can be concluded that these estimations should be considered valid and reliable and that cipa was immedi- ately and consistently implemented in the majority of the public schools since . how many public libraries have implemented cipa? in , . percent of public libraries were receiving e-rate discounts, and . percent said they would not apply for e-rate discounts if cipa was upheld. since the supreme court decision upholding cipa, the num- ber of libraries complying with cipa has increased, as table . implementation of cipa in public schools year access (%) filtering (%) information technology and libraries | march have the number of libraries not applying for e-rate funds to avoid complying with cipa. however, unlike schools, there is no exact count of how many libraries have filtered internet access. in many cases, the libraries themselves do not filter, but a state library, library con- sortium, or local or state government system of which they are a part filters access from beyond the walls of the library. in some of these cases, the library staff may not even be aware that such filtering is occurring. a number of state and local governments have also passed their own laws to encourage or require all libraries in the state to filter internet access regardless of e-rate or lsta funds. in , . percent of public libraries were filtering access within the library as a result of directly receiving e-rate funding. furthermore, . percent of libraries were receiving e-rate funding as a part of another orga- nization, meaning that these libraries also would need to comply with cipa’s requirements. as such, the number of public libraries filtering access is now at least . percent, but the number will likely be higher as a result of state and local laws requiring libraries to filter as well as other reasons libraries have implemented filters. in contrast, among libraries not receiving e-rate funds, the number of libraries now not applying for e-rate inten- tionally to avoid the cipa requirements is . percent. while it is not possible to identify an exact number of public libraries that filter access, it is clear that libraries overall have far lower levels of filtering than the per- cent of public schools that filter access. e-rate and other program issues the administration of the e-rate program has not occurred without controversy. throughout the course of the program, many applicants for and recipients of the funding have found the program structure to be obtuse, the application process to be complicated and time con- suming, and the administration of the decision-making process to be slow. as a result, many schools and librar- ies find it difficult to plan ahead for budgeting purposes, not knowing how much funding they will receive or when they will receive it. there also have been larger difficulties for the program. following revelations about the uses of some e-rate awards, the fcc suspended the program from august to december to impose new accounting and spending rules for the funds, delaying the distribution of over $ billion in funding to libraries and schools. news inves- tigations had discovered that certain school systems were using e-rate funds to purchase more technology than they needed or could afford to maintain, and some school systems failed to ever use technology they had acquired. while the administration of the e-rate program has been comparatively smooth since, the temporary suspension of the program caused serious short-term problems for, and left a sense of distrust of, the program among many recipients. filtering issues during the s, many types of software filtering prod- ucts became available to consumers, including server- side filtering products (using a list of server-selected blocked urls that may or may not be disclosed to the user), client-side filtering (controlling the blocking of specific content with a user password), text-based content-analysis filtering (removing illicit content of a website using real-time analysis), monitoring and time- limiting technologies (tracking a child’s online activi- ties and limiting the amount of time he or she spends online), and age-verification systems (allowing access to webpages by passwords issued by a third party to an adult). but because filtering software companies make the decisions about how the products work, content and collection decisions for electronic resources in schools and public libraries have been taken out of the hands of librarians, teachers, and local communities and placed in the trust of proprietary software products. some filtering programs also have specific political agendas, which many organizations that purchase them are not aware of. in a study of over one million pages, for every webpage blocked by a filter as advertised by the software vendor, one or more pages were blocked inappropriately, while many of the criteria used by the filtering products go beyond the criteria enumerated in cipa. filters have significant rates of inappropriately block- ing materials, meaning that filters misidentify harmless materials as suspect and prevent access to harmless items (e.g., one filter blocked access to the declaration of independence and the constitution). furthermore, when libraries install filters to comply with cipa, in many instances the filters will frequently be blocking text as well as images, and (depending on the type of filter- ing product employed) filters may be blocking access to entire websites or even all the sites from certain internet service providers. as such, the current state of filtering technology will create the practical effect of cipa restrict- ing access to far more than just certain types of images in many schools and libraries. n differences in the perceived value of cipa and filtering based on the available data, there clearly is a sizeable contrast in the levels of implementation of cipa between one law with two outcomes | jaeger and yan schools and libraries. this difference raises a number of questions: for what reasons has cipa been much more widely implemented in schools? is this issue mainly value driven, dollar driven, both, or neither in these two public institutions? why are these two institutions so dif- ferent regarding cipa implementation while they share many social and educational similarities? reasons for nationwide full implementation in schools there are various reasons—from financial, population, social, and management issues to computer and internet availability—that have driven the rapid and compre- hensive implementation of filters in public schools. first, public schools have to implement cipa because of societal pressures and the lobbying of parents to ensure students’ internet safety. almost all users of computers in schools are minors, the most vulnerable groups for internet crimes and child pornography. public schools in america have been the focus of public attention and scru- tiny for years, and the political and social responsibility of public schools for children’s internet safety is huge. as a result, society has decided these students should be most strongly protected, and cipa was implemented immediately and most widely at schools. second, in contrast to public libraries (which average slightly less than eleven computers per library outlet), the typical number of computers in public schools ranges from one hundred to five hundred, which are needed to meet the needs of students and teachers for daily learning and teaching. since the number of computers is quite large, the financial incentives of e-rate funding are substantial and critical to the operation of the schools. this situation provides administrators in schools and school districts with the incentive to make decisions to implement cipa as quickly and extensively as possible. furthermore, the amount of money that e-rate provides for schools in terms of technology is astounding. as was noted earlier, schools received over $ billion from to alone. schools likely would not be able to provide the necessary computers for students and teachers with- out the e-rate funds. third, the actual implementation procedure differs in schools and libraries: schools are certified under the supervision of the local educational agencies such as school districts and state departments of education; libraries are certified within a library organization under the supervision of the imls. in other words, the cer- tification process at schools is directly and effectively controlled by school districts and state departments of education, following the same fundamental values of protecting children. the resistance to cipa in schools has been very small in comparison to libraries. the primary concern raised has been the issue of educational equality. concerns have been raised that filters in schools may create two classes of students—ones with only filtered access at school and ones who also can get unfiltered access at home. reasons for more limited implementation in libraries in public libraries, the reasons for implementing cipa are similar to those of public schools in many ways. public libraries provide an average of . computers in each of the approximately seven thousand public libraries in the united states, which is a lot of technology that needs to be supported. the e-rate and lsta funds are vital to many libraries in the provision of computers and the internet. furthermore, with limited alternative sources of funding, the e-rate and lsta funds are hard to replace if they are not available. given that the public libraries have become the guarantor of public access to comput- ing and the internet, libraries have to find ways to ensure that patrons can access the internet. libraries also have to be concerned about protect- ing and providing a safe environment for younger patrons. while libraries serve patrons of all ages, one of the key social expectations of libraries is the provision of educational materials for children and young adults. children’s sections of libraries almost always have com- puters in them. much of the content blocked by filters is of little or no education value. as such, “defending unfil- tered internet access was quite different from defending catcher in the rye.” nevertheless, many libraries have fought against the filtering requirements of cipa because they believe that it violates the principles of librarianship or for a number of other reasons. in , . percent of public libraries refused to apply for e-rate or lsta funds specifically to avoid cipa requirements, a substantial increase from the . percent of libraries that did not apply for e-rate because of cipa in . as a result of defending patron’s rights to free access, the libraries that are not applying for e-rate funds because of the requirements of cipa are being forced to turn down the chance for fund- ing to help pay for internet access in order to preserve community access to the internet. because many librar- ies feel that they cannot apply for e-rate funds, local and regional discrepancies are occurring in the levels of internet access that are available to patrons of public libraries in different parts of the country. for adult patrons who wish to access material on computers with filters, cipa states that the library has the option of disabling the filters for “bona fide research or other lawful purposes” when adult patrons request such disabling. the law does not require libraries to information technology and libraries | march disable the filters for adult patrons, and the criteria for disabling of filters do not have a set definition in the law. the potential problems in the process of having the filters disabled are many and significant, including librarians not allowing the filters to be turned off, librarians not knowing how to turn the filters off, the filtering software being too complicated to turn off without injuring the performance of the workstation in other applications, or the filtering software being unable to be turned off in a reasonable amount of time. it has been estimated that approximately million low-income individuals rely on public libraries to access online information because they lack internet access at home or work. the e-rate and lsta programs have helped to make public libraries a trusted community source of internet access, with the public library being the only source of free public internet access available to all community residents in nearly percent of communities in the united states. therefore usage of computers and the internet in public libraries has continued to grow at a very fast pace over the past ten years. thus public librar- ies are torn between the values of providing safe access for younger patrons and broad access for adult patrons who may have no other means of accessing the internet. n cipa, public policy, and further research while the diverse implementations, effects, and levels of acceptance of cipa across schools and libraries dem- onstrate the wide range of potential ramifications of the law, surprisingly little consideration is given to major assumptions in the law, including the appropriateness of the requirements to different age groups and the nature of information on the internet. cipa treats all users as if they are the same level of maturity and need the same level of protection as a small child, as evidenced by the require- ment that all computers in a library or school have filters regardless of whether children use a particular computer. in reality, children and adults interact in different social, physical, and cognitive ways with computers because of different developmental processes. cipa fails to recognize that children as individual users are active processors of information and that children of different ages are going to be affected in divergent ways by filtering programs. younger children benefit from more restrictive filters while older children benefit from less restrictive filters. moreover, filtering can be compli- mented by encouragement of frequent positive internet usage and informal instruction to encourage positive use. finally, children of all ages need a better understanding of the structure of the internet to encourage appropriate caution in terms of online safety. the internet represents a new social and cultural environment in which users simultaneously are affected by the social environment and also construct that environment with other users. cipa also is based on fundamental misconceptions about information on the internet. the supreme court’s decision upholding cipa represents several of these mis- conceptions, adopting an attitude that ‘we know what is best for you’ in terms of the information that citizens should be allowed to access. it assumes that schools and libraries select printed materials out of a desire to protect and censor rather than recognizing the basic reality that only a small number of print materials can be afforded by any school or library. the internet frees schools and libraries from many of these costs. furthermore, the court assumes that libraries should censor the internet as well, ultimately upholding the same level of access to information for adult patrons and librarians in public libraries as students in public schools. these two major unexamined assumptions in the law certainly have played a part in the difficulty of implementing cipa and in the resistance to the law. and this does not even address the problems of assuming that public libraries and public schools can be treated interchangeably in crafting legislation. these problem- atic assumptions point to a significantly larger issue: in trying to deal with the new situations created by the internet and related technology, the federal government has significantly increased the attention paid to informa- tion policy. over the past few years, government laws and standards related to information have begun to more clearly relate to social aspects of information technolo- gies such as the filtering requirements of cipa. but the social, economic, and political ramifications for decisions about information policy are often woefully underexam- ined in the development of legislation. this paper has documented that many of the reasons for and statistics about cipa implementation are avail- able by bringing together information from different social institutions. the biggest questions about cipa are about the societal effects of the policy decisions: n has cipa changed the education and information- provision roles of libraries and schools? n has cipa changed the social expectations for libraries and schools? n have adult patron information behaviors changed in libraries? n have minor patron information behaviors changed in libraries? n have student information behaviors changed in school? n how has cipa changed the management of librar- ies and schools? n will congress view cipa as successful enough to merit using libraries and schools as the means of enforcing other legislation? one law with two outcomes | jaeger and yan but these social and administrative concerns are not the only major research questions raised by the imple- mentation of cipa. future research about cipa not only needs to focus on the individual, institutional, and social effects of the law. it must explore the lessons that cipa can provide to the process of creating and implementing information policies with significant societal implications. the most significant research issues related to cipa may be the ones that help illuminate how to improve the legislative process to better account for the potential consequences of regulating information while the legislation is still being developed. such cross-disciplinary analyses would be of great value as information becomes the center of an increasing amount of legislation, and the effects of this legislation have continually wider consequences for the flow of information through society. it could also be of great benefit to public schools and libraries, which, if cipa is any indication, may play a large role in future legislation about public internet access. references . children’s internet protection act (cipa), public law - . . united states v. american library association, u.s. ( ). . american library association, libraries connect communi- ties: public library funding & technology access study – (chicago: ala, ); paul t. jaeger, john carlo bertot, and charles r. mcclure, “the effects of the children’s internet protection act (cipa) in public libraries and its implications for research: a statistical, policy, and legal analysis,” journal of the american society for information science and technology , no. ( ): – ; paul t. jaeger et al., “public libraries and internet access across the united states: a comparison by state from to ,” information technology and librar- ies , no. ( ): – ; paul t. jaeger et al., “cipa: decisions, implementation, and impacts,” public libraries , no. ( ): – ; zheng yan, “limited knowledge and limited resources: children’s and adolescents’ understanding of the internet,” journal of applied developmental psychology (forthcoming); zheng yan, “differences in basic knowledge and perceived education of internet safety between high school and undergraduate students: do high school students really benefit from the children’s internet protection act?” journal of applied develop- mental psychology (forthcoming); zheng yan, “what influences children’s and adolescents’ understanding of the complexity of the internet?,” developmental psychology ( ): – . . martha m. mccarthy, “filtering the internet: the chil- dren’s internet protection act,” educational horizons , no, (winter ): . . federal communications commission, in the matter of federal–state joint board on universal service: children’s internet protection act, fcc order - (washington, d.c.: ). . cipa. . roth v. united states, u.s. ( ). . miller v. california, u.s. ( ). . roth v. united states. . cipa. . cipa. . telecommunications act of , public law - (feb. , ). . paul t. jaeger, charles r. mcclure, and john carlo ber- tot, “the e-rate program and libraries and library consortia, – : trends and issues,” information technology & libraries , no. ( ): – . . ibid. . ibid. . american library association, “u.s. supreme court arguments on cipa expected in late winter or early spring,” press release, nov. , , www.ala.org/ala/aboutala/hqops/ pio/pressreleasesbucket/ussupremecourt.cfm (accessed may , ). . kelly rodden, “the children’s internet protection act in public schools: the government stepping on parents’ toes?” fordham law review ( ): – . . john carlo bertot, paul t. jaeger, and charles r. mcclure, “public libraries and the internet : issues, implications, and expectations,” library & information science research ( ): – ; charles r. mcclure, paul t. jaeger, and john carlo bertot, “the looming infrastructure plateau?: space, funding, connection speed, and the ability of public libraries to meet the demand for free internet access,” first monday , no. ( ), www.uic.edu/htbin/cgiwrap/bin/ojs/index.php/fm/ article/view/ / (accessed may , ). . mccarthy, “filtering the internet.” . leigh s. estabrook and edward lakner, “managing inter- net access: results of a national survey,” american libraries , no. ( ): – . . alberta davis comer, “studying indiana public librar- ies’ usage of internet filters,” computers in libraries (june ): – ; thomas m. reddick, “building and running a collabora- tive internet filter is akin to a kansas barn raising,” computers in libraries , no. ( ): – . . communications decency act of , public law - (feb. , ). . child online protection act (copa), public law - (oct. , ). . united states v. american library association. . r. trevor hall and ed carter, “examining the constitu- tionality of internet filtering in public schools: a u.s. perspec- tive,” education & the law , no. ( ): – ; mccarthy “filtering the internet.” . library services and technology act, public law - (sept. , ). . john wells and laurie lewis, internet access in u.s. public schools and classrooms: – , special report prepared at the request of the national center for education statistics, nov. . . american library association, libraries connect commu- nities; john carlo bertot, charles r. mcclure, and paul t. jaeger, “the impacts of free public internet access on public library patrons and communities,” library quarterly , no. ( ): – ; jaeger et al., “cipa.” . wells and lewis, internet access in u.s. public schools and classrooms. information technology and libraries | march . ibid. . jaeger, mcclure, and bertot, “the e-rate program and libraries and library consortia.” . jaeger et al., “cipa.” . american library association, libraries connect commu- nities. . ibid. . ibid. . jaeger, mcclure, and bertot, “the e-rate program and libraries and library consortia.” . ibid. . norman oder, “$ million in e-rate funds suspended: delays caused as fcc requires new accounting standards,” library journal , no. ( ): ; debra lau whelan, “e-rate funding still up in the air: schools, libraries left in the dark about discounted funds for internet services,” school library journal , no. ( ): . . ken foskett and paul donsky, “hard eye on city schools’ hardware,” atlanta journal-constitution, may , ; ken fos- kett and jeff nesmith, “wired for waste: abuses tarnish e-rate program,” atlanta journal-constitution, may , . . jaeger, mcclure, and bertot, “the e-rate program and libraries and library consortia.” . department of commerce, national telecommunication and information administration, children’s internet protection act: study of technology protection measures in section , report to congress (washington, d.c.: ). . mccarthy, “filtering the internet.” . paul t. jaeger and charles r. mcclure, “potential legal challenges to the application of the children’s internet protec- tion act (cipa) in public libraries: strategies and issues,” first monday , no. ( ), www.firstmonday.org/issues/issue _ / jaeger/index.html (accessed may , ). . electronic frontier foundation, internet blocking in public schools (washington, d.c.: ), http://w .eff.org/censor ship/censorware/net_block_report (accessed may , ). . adam horowitz, “the constitutionality of the children’s internet protection act,” st. thomas law review , no. ( ): – . . tanessa cabe, “regulation of speech on the internet: fourth time’s the charm?” media law and policy ( ): – ; adam goldstein, “like a sieve: the child internet pro- tection act and ineffective filters in libraries,” fordham intel- lectual property, media, and entertainment law journal ( ): – ; horowitz, “the constitutionality of the children’s internet protection act”; marilyn j. maloney and julia morgan, “rock and a hard place: the public library’s dilemma in pro- viding access to legal materials on the internet while restrict- ing access to illegal materials,” hamline law review , no. ( ): – ; mary minow, “filters and the public library: a legal and policy analysis,” first monday , no. ( ), www .firstmonday.org/issues/issue _ /minnow (accessed may , ); richard j. peltz, “use ‘the filter you were born with’: the unconstitutionality of mandatory internet filtering for adult patrons of public libraries,” washington law review , no. ( ): – . . mccarthy, “filtering the internet.” . john carlo bertot et al., “public access computing and internet access in public libraries: the role of public libraries in e-government and emergency situations,” first monday , no. ( ), www.firstmonday.org/issues/issue _ /bertot (accessed may , ); john carlo bertot et al., “drafted: i want you to deliver e-government,” library journal , no. ( ): – ; paul t. jaeger and kenneth r. fleischmann, “public libraries, values, trust, and e-government,” informa- tion technology and libraries , no. ( ): – . . doug johnson, “maintaining intellectual freedom in a filtered world,” learning & leading with technology , no. (may ): . . bertot, mcclure, and jaeger, “the impacts of free public internet access on public library patrons and communities.” . jaeger et al., “public libraries and internet access across the united states.” . paul t. jaeger et al., “the policy implications of internet connectivity in public libraries,” government information quar- terly , no. ( ): – . . goldstein, “like a sieve.” . bertot, mcclure, and jaeger, “the impacts of free public internet access on public library patrons and communities”; jaeger and fleischmann, “public libraries, values, trust, and e-government.“ . bertot, jaeger, and mcclure, “public libraries and the internet ”; charles r. mcclure et al., “funding and expen- ditures related to internet access in public libraries,” informa- tion technology & libraries (forthcoming). . zheng yan and kurt w. fischer, “how children and adults learn to use computers: a developmental approach,” new directions for child and adolescent development ( ): – . . zheng yan, “age differences in children’s understand- ing of the complexity of the internet,” journal of applied devel- opmental psychology ( ): – ; yan, “limited knowledge and limited resources”; yan, “differences in basic knowledge and perceived education of internet safety”; yan, “what influ- ences children’s and adolescents’ understanding of the com- plexity of the internet?” . patricia greenfield and zheng yan, “children, adoles- cents, and the internet: a new field of inquiry in developmen- tal psychology,” developmental psychology ( ): – . . john n. gathegi, “the public library as a public forum: the (de)evolution of a legal doctrine,” library quarterly ( ): . . sandra braman, “where has media policy gone? defin- ing the field in the st century,” communication law and policy , no. ( ): – ; sandra braman, change of state: informa- tion, policy, & power (cambridge, mass.: mit pr., ); charles r. mcclure and paul t. jaeger, “government information policy research: importance, approaches, and realities,” library & information science research ( ): – ; milton mueller, christiane page, and brendan kuerbis, “civil society and the shaping of communication-information policy: four decades of advocacy,” information society , no. ( ): – . . paul t. jaeger, “information policy, information access, and democratic participation: the national and international implications of the bush administration’s information politics,” government information quarterly ( ): – . . mcclure and jaeger, “government information policy research.” information technology and libraries | march mathew j. miles and scott j. bergstrom classification of library resources by subject on the library website: is there an optimal number of subject labels? the number of labels used to organize resources by subject varies greatly among library websites. some librarians choose very short lists of labels while others choose much longer lists. we conducted a study with students and staff to try to answer the following question: what is the effect of the number of labels in a list on response time to research questions? what we found is that response time increases gradually as the number of the items in the list grow until the list size reaches approximately fifty items. at that point, response time increases significantly. no asso- ciation between response time and relevance was found. i t is clear that academic librarians face a daunting task drawing users to their library’s web presence. “nearly three-quarters ( %) of college students say they use the internet more than the library, while only % said they use the library more than the internet for informa- tion searching.” improving the usability of the library websites therefore should be a primary concern for librar- ians. one feature common to most library websites is a list of resources organized by subject. libraries seem to use similar subject labels in their categorization of resources. however, the number of subject labels varies greatly. some use as few as five subject labels while others use more than one hundred. in this study we address the following ques- tion: what is the effect of the number of subject labels in a list on response times to research questions? n literature review mcgillis and toms conducted a performance test in which users were asked to find a database by navigating through a library website. they found that participants “had difficulties in choosing from the categories on the home page and, subsequently, in figuring out which data- base to select.” a review of relevant research literature yielded a number of theses and dissertations in which the authors compared the usability of different library websites. jeng in particular analyzed a great deal of the usability testing published concerning the digital library. the following are some of the points she summarized that were highly relevant to our study: n user “lostness”: users did not understand the structure of the digital library. n ambiguity of terminology: problems with wording accounted for percent of usability problems. n finding periodical articles and subject-specific databases was a challenge for users. a significant body of research not specific to libraries provides a useful context for the present research. miller’s landmark study regarding the capacity of human short- term memory showed as a rule that the span of immedi- ate memory is about ± items. sometimes this finding is misapplied to suggest that menus with more than nine subject labels should never be used on a webpage. subsequent research has shown that “chunking,” which is the process of organizing items into “a collection of ele- ments having strong associations with one another, but weak associations with elements within other chunks,” allows human short-term memory to handle a far larger set of items at a time. larson and czerwinski provide important insights into menuing structures. for example, increasing the depth (the number of levels) of a menu harms search performance on the web. they also state that “as you increase breadth and/or depth, reaction time, error rates, and perceived complexity will all increase.” however, they concluded that a “medium condition of breadth and depth outperformed the broadest, shallow web structure overall.” this finding is somewhat contrary to a previous study by snowberry, parkinson, and sisson, who found that when testing structures of , , , ( means two menu items per level, six levels deep), the structure grouped into categories proved to be advantageous in both speed and accuracy. larson and czerwinksi rec- ommended that “as a general principle, the depth of a tree structure should be minimized by providing broad menus of up to eight or nine items each.” zaphiris also corroborated that previous research con- cerning depth and breadth of the tree structure was true for the web. the deeper the tree structure, the slower the user performance. he also found that response times for expandable menus are on average percent longer than sequential menus. both the research and current practices are clear concerning the efficacy of hierarchical menu structures. thus it was not a focus of our research. the focus instead was on a single-level menu and how the number and characteristics of subject labels would affect search response times. n background in preparation for this study, library subject lists were col- lected from a set of thirty library websites in the united mathew j. miles (milesm@byui.edu) is systems librarian and scott j. bergstrom (bergstroms@byui.edu) is director of institutional research at brigham young university–idaho in rexburg. classification of library resources by subject on the library website | miles and bergstrom states, canada, and the united kingdom. we selected twelve lists from these websites that were representative of the entire group and that varied in size from small to large. to render some of these lists more usable, we made slight modifications. there were many similarities between label names. n research design participants were randomly assigned to one of twelve experimental groups. each experimental group would be shown one of the twelve lists that were selected for use in this study. roughly percent of the participants were students. the remaining percent of the participants were full-time employees who worked in these same departments. the twelve lists ranged in number of labels from five to seventy-two: n group a: subject labels n group b: subject labels n group c: subject labels n group d: subject labels n group e : subject labels n group f: subject labels n group g: subject labels n group h: subject labels n group i: subject labels n group j: subject labels n group k: subject labels n group l: subject labels each participant was asked to select a subject label from a list in response to eleven different research ques- tions. the questions are listed below: . which category would most likely have informa- tion about modern graphical design? . which category would most likely have informa- tion about the aztec empire of ancient mexico? . which category would most likely have informa- tion about the effects of standardized testing on high school classroom teaching? . which category would most likely have informa- tion on skateboarding? . which category would most likely have informa- tion on repetitive stress injuries? . which category would most likely have informa- tion about the french revolution? . which category would most likely have informa- tion concerning walmart’s marketing strategy? . which category would most likely have information on the reintroduction of wolves into yellowstone park? . which category would most likely have informa- tion about the effects of increased use of nuclear power on the price of natural gas? . which category would most likely have informa- tion on the electoral college? . which category would most likely have informa- tion on the philosopher emmanuel kant? the questions were designed to represent a variety of subject areas that library patrons might pursue. each sub- ject list was printed on a white sheet of paper in alphabetical order in a single column, or double columns when needed. we did not attempt to test the subject lists in the context of any web design. we were more interested in observing the effect of the number of labels in a list on response time inde- pendent of any web design. each participant was asked the same eleven questions in the same order. the order of ques- tions was fixed because we were not interested in testing for the effect of order and wanted a uniform treatment, thereby not introducing extraneous variance into the results. for each question, the participant was asked to select a label from the subject list under which they would expect to find a resource that would best provide information to answer the question. participants were also instructed to select only a single label, even if they could think of more than one label as a possible answer. participants were encour- aged to ask for clarification if they did not fully understand the question being asked. recording of response times did not begin until clarification of the question had been given. response times were recorded unbeknownst to the partici- pant. if the participant was simply unable to make a selec- tion, that was also recorded. two people administered the exercise. one recorded response times; the other asked the questions and recorded label selections. relevance rankings were calculated for each possible combination of labels within a subject list for each ques- tion. for example, if a subject list consisted of five labels, for each question there were five possible answers. two library professionals—one with humanities expertise, the other with sciences expertise—assigned a relevance rank- ing to every possible combination of question and labels within a subject list. the rankings were then averaged for each question–label combination. n results the analysis of the data was undertaken to determine whether the average response times of participants, adjusted by the different levels of relevance in the subject list labels that prevailed for a given question, were signifi- cantly different across the different lists. in other words, would the response times of participants using a particu- lar list, for whom the labels in the list were highly relevant information technology and libraries | march to the question, be different from students using the other lists for whom the labels in the list were also highly relevant to the question? a separate univariate general linear model analysis was conducted for each of the eleven questions. the analyses were conducted separately because each ques- tion represented a unique search domain. the univariate general linear model pro- vided a technique for testing whether the average response times associated with the different lists were significantly dif- ferent from each other. this technique also allowed for the inclusion of a cova- riate—relevance of the subject list labels to the question—to determine whether response times at an equivalent level of relevance was different across lists. in the analysis model, the depen- dent variable was response time, defined as the time needed to select a subject list label. the covariate was relevance, defined as the perceived match between a label and the question. for example, a label of “economics” would be assessed as highly relevant to the question, what is the current unemployment rate? the same label would be assessed as not relevant for the question, what are the names of four moons of saturn? the main factor in the model was the actual list being presented to the participant. there were twelve lists used in this study. the statistical model can be summarized as follows: response time = list + relevance + (list × relevance) + error the general linear model required that the following conditions be met: first, data must come from a ran- dom sample from a normal population. second, all vari- ances with each of the groupings are the same (i.e., they have homoscedasticity). an examination of whether these assumptions were met revealed problems both with nor- mality and with homoscedasticity. a common technique— logarithmic transformation—was employed to resolve these problems. accordingly, response-time data were all converted to common logarithms. an examination of assumptions with the transformed data showed that all questions but three met the required conditions. the three . . . . . . . . avg log performance trend figure . the overall average of average search times for the eight questions for all experimental groups (i.e., lists) questions ( , , and ) were excluded from subsequent analysis. n conclusions the series of graphs in the appendix show the average response times, adjusted for relevance, for eight of the eleven questions for all twelve lists (i.e., experimental groups). three of the eleven questions were excluded from the analysis because of heteroscedascity. an inspec- tion of these graphs shows no consistent pattern in response time as the number of the items in the lists increase. essentially, this means that, for any given level of relevance, the number of items of the list does not affect response time significantly. it seems that for a single ques- tion, characteristics of the categories themselves are more important than the quantity of categories in the list. the response times using a subject list with twenty-eight labels is similar to the response times using a list of six labels. a statistical comparison of the mean response time for each classification of library resources by subject on the library website | miles and bergstrom group with that of each of the other groups for each of the questions largely confirms this. there were very few statistically significant different comparisons. the spikes and valleys of the graphs in the appendix are generally not significantly different. however, when the average response time associated with all lists is combined into an overall average from all eight questions, a somewhat clearer picture emerges (see figure ). response times increase gradually as the number of the items in the list increase until the list size reaches approximately fifty items. at that point, response time increases significantly. no association was found between response time and relevance. a fast response time did not necessarily yield a relevant response, nor did a slow response time yield an irrelevant response. n observations we observed that there were two basic patterns exhibited when participants made selections. the first pattern was the quick selection—participants easily made a selection after performing an initial scan of the available labels. nevertheless, a quick selection did not always mean a relevant selection. the second pattern was the delayed selection. if participants were unable to make a selection after the initial scan of items, they would hesitate as they struggled to determine how the question might be reclas- sified to make one of the labels fit. we did not have access to a high-tech lab, so we were unable to track eye move- ment, but it appeared that the participants began scan- ning up and down the list of available items in an attempt to make a selection. the delayed selection seemed to be a combination of two problems: first, none of the avail- able labels seemed to fit. second, the delay in scanning increased as the list grew larger. it’s possible that once the list becomes large enough, scanning begins to slow the selection process. a delayed selection did not necessarily yield an irrelevant selection. the label names themselves did not seem to be a significant factor affecting user performance. we did test three lists, each with nine items and each having differ- ent labels, and response times were similar for the three lists. a future study might compare a more extensive number of lists with the same number of items with different labels to see if label names have an effect on response time. this is a particular challenge to librarians in classifying the digital library, since they must come up with a few labels to classify all possible subjects. creating eleven questions to span a broad range of subjects is also a possible weakness of the study. we had to throw out three questions that violated the assump- tions of the statistical model. we tried our best to select questions that would represent the broad subject areas of science, arts, and general interest. we also attempted to vary the difficulty of the questions. a different set of questions may yield different results. references . steve jones, the internet goes to college, ed. mary madden (washington, d.c.: pew internet and american life project, ): , www.pewinternet.org/pdfs/pip_college_report.pdf (accessed mar. , ). . louise mcgillis and elaine g. toms, “usability of the academic library web site: implications for design,” college & research libraries , no. ( ): . . judy h. jeng, “usability of the digital library: an evalu- ation model” (phd diss., rutgers university, new brunswick, new jersey): – . . george a. miller, “the magical number seven plus or minus two: some limits on our capacity for processing infor- mation,” psychological review , no. ( ): – . . fernand gobet et al., “chunking mechanisms in human learning,” trends in cognitive sciences , no. ( ): – . . kevin larson and mary czerwinski, “web page design: implications of memory, structure and scent for informa- tion retrieval” (los angeles: acm/addison-wesley, ): , http://doi.acm.org/ . / . (accessed nov. , ). . ibid. . kathleen snowberry, mary parkinson, and norwood sis- son, “computer display menus,” ergonomics , no ( ): . . larson and czerwinski, “web page design,” . . panayiotis g. zaphiris, “depth vs. breath in the arrange- ment of web links,” www.soi.city.ac.uk/~zaphiri/papers/hfes .pdf (accessed nov. , ). . panayiotis g. zaphiris, ben shneiderman, and kent l. norman, “expandable indexes versus sequential menus for searching hierarchies on the world wide web,” http:// citeseer.ist.psu.edu/rd/ % c % c % c . % cdow nload/http://coblitz.codeen.org: /citeseer.ist.psu.edu/ cache/papers/cs/ /http:zszzszagrino.orgzszpzaphiriz szpaperszszexpandableindexes.pdf/zaphiris expandable.pdf (accessed nov. , ). information technology and libraries | march appendix. response times by question by group . . . . . . . gr p a ( it em s) gr p e ( it em s) gr p f ( it em s) gr p b ( it em s) gr p c ( it em s) gr p h ( it em s) gr p g ( ite m s) gr p d ( ite m s) gr p j ( ite m s) gr p i ( ite m s) gr p k ( ite m s) gr p l ( ite m s) . . . . . . . gr p a ( it em s) gr p e ( it em s) gr p f ( it em s) gr p b ( it em s) gr p c ( it em s) gr p h ( it em s) gr p g ( ite m s) gr p d ( ite m s) gr p j ( ite m s) gr p i ( ite m s) gr p k ( ite m s) gr p l ( ite m s) . . . . . . . gr p a ( it em s) gr p e ( it em s) gr p f ( it em s) gr p b ( it em s) gr p c ( it em s) gr p h ( it em s) gr p g ( ite m s) gr p d ( ite m s) gr p j ( ite m s) gr p i ( ite m s) gr p k ( ite m s) gr p l ( ite m s) . . . . . . . . gr p a ( it em s) gr p e ( it em s) gr p f ( it em s) gr p b ( it em s) gr p c ( it em s) gr p h ( it em s) gr p g ( ite m s) gr p d ( ite m s) gr p j ( ite m s) gr p i ( ite m s) gr p k ( ite m s) gr p l ( ite m s) . . . . . . . gr p a ( it em s) gr p e ( it em s) gr p f ( it em s) gr p b ( it em s) gr p c ( it em s) gr p h ( it em s) gr p g ( ite m s) gr p d ( ite m s) gr p j ( ite m s) gr p i ( ite m s) gr p k ( ite m s) gr p l ( ite m s) . . . . . . . gr p a ( it em s) gr p e ( it em s) gr p f ( it em s) gr p b ( it em s) gr p c ( it em s) gr p h ( it em s) gr p g ( ite m s) gr p d ( ite m s) gr p j ( ite m s) gr p i ( ite m s) gr p k ( ite m s) gr p l ( ite m s) . . . . . . . . . gr p a ( it em s) gr p b ( it em s) gr p c ( it em s) gr p d ( ite m s) gr p e ( it em s) gr p f ( it em s) gr p g ( ite m s) gr p h ( it em s) gr p i ( ite m s) gr p j ( ite m s) gr p k ( ite m s) gr p l ( ite m s) . . . . . . . . . . gr p a ( it em s) gr p e ( it em s) gr p f ( it em s) gr p b ( it em s) gr p c ( it em s) gr p h ( it em s) gr p g ( ite m s) gr p d ( ite m s) gr p j ( ite m s) gr p i ( ite m s) gr p k ( ite m s) gr p l ( ite m s) question question question question question question question question laneconnex | ketchell et al. laneconnex: an integrated biomedical digital library interface debra s. ketchell, ryan max steinberg, charles yates, and heidi a. heilemann this paper describes one approach to creating a search application that unlocks heterogeneous content stores and incorporates integrative functionality of web search engines. laneconnex is a search interface that identifies journals, books, databases, calculators, bioinformatics tools, help information, and search hits from more than three hundred full-text heterogeneous clinical and biore- search sources. the user interface is a simple query box. results are ranked by relevance with options for filtering by content type or expanding to the next most likely set. the system is built using component-oriented program- ming design. the underlying architecture is built on apache cocoon, java servlets, xml/xslt, sql, and javascript. the system has proven reliable in production, reduced user time spent finding information on the site, and maximized the institutional investment in licensed resources. m ost biomedical libraries separate searching for resources held locally from external database searching, requiring clinicians and researchers to know which interface to use to find a specific type of information. google, amazon, and other web search engines have shaped user behavior and expectations. users expect a simple query box with results returned from a broad array of content ranked or categorized appropriately with direct links to content, whether it is an html page, a pdf document, a streaming video, or an image. biomedical libraries have transitioned to digital journals and reference sources, adopted openurl link resolvers, and created institutional repositories. however, students, clinicians, and researchers are hindered from maximizing this content because of proprietary and het- erogeneous systems. a strategic challenge for biomedical libraries is to create a unified search for a broad spectrum of licensed, open-access, and institutional content. n background studies show that students and researchers will use the search path of least cognitive resistance. ease and speed are the most important factors for using a particular search engine. a university of california report found that academic users want one search tool to cover a wide information universe, multiple formats, full-text avail- ability to move seamlessly to the item itself, intelligent assistance and spelling correction, results sorted in order of relevance, help navigating large retrievals by logical subsetting and customization, and seamless access any- time, anywhere. studies of clinicians in the patient-care environment have documented that effort is the most important factor in whether a patient-care question is pursued. for researchers, finding and using the best bio- informatics tool is an elusive problem. in , the lane medical library and knowledge management center (lane) at the stanford university medical center provided access to an expansive array of licensed, institutional, and open-access digital content in support of research, patient care, and education. like most of its peers, lane users were required to use scores of different interfaces to search external databases and find digital resources. we created a local metasearch application for clinical reference content, but it did not integrate result sets from disparate resources. a review of federated-search software in the marketplace found that products were either slow or they limited retrieval when faced with a broad spectrum of biomedical content. we decided to build on our existing application architecture to create a fast and unified interface. a detailed analysis of lane website-usage logs was conducted before embarking on the creation of the new search application. key points of user failure in the exist- ing search options were spelling errors that could easily be corrected to avoid zero results; lack of sufficient intui- tive options to move forward from a zero-results search or change topics without backtracking; lack of use of existing genre or role searches; confusion about when to use the resource, openurl resolver, or pubmed search to find a known item; and results that were cognitively difficult to navigate. studies of the web search engine and the pubmed search log concurred with our usage- log analysis: a single term search is the most common, with three words maximum entered by typical users. a pubmed study found that percent of user queries were for known items rather than for a general subject, con- firming our own log analysis findings that the majority of searches were for a particular source item. search-term analysis revealed that many of our users were entering partial article citations (e.g., author, date) in any query debra s. ketchell (debra.ketchell@gmail.com) is the for- mer associate dean for knowledge management and library director; ryan max steinberg (ryan.max.steinberg@stanford .edu) is the knowledge integration programmer/architect; charles yates (charles.yates@stanford.edu) is the systems software developer; and heidi a. heilemann (heidi.heilemann@stanford .edu) is the former director for research & instruction and cur- rent associate dean for knowledge management and library director at the lane medical library & knowledge management center, information resources & technology, stanford university school of medicine, stanford, california. information technology and libraries | march box expecting that article databases would be searched concurrently with the resource database. our displayed results were sorted alphabetically, and each version of an item was displayed separately. for the user, this meant a cluttered list with redundant title information that increased their cognitive effort to find meaningful items. overall, users were confronted with too many choices upfront and too few options after retrieving results. focus groups of faculty and students were conducted in . attendees wanted local information integrated into the proposed single search. local information included content such as how-to information, expertise, seminars, grand rounds, core lab resources, drug formulary, patient handouts, and clinical calculators. most of this content is restricted to the stanford user population. users consis- tently described their need for a simple search interface that was fast and customized to the stanford environ- ment. in late , we embarked on a project to design a search application that would address both existing points of failure in the current system and meet the expressed need for a comprehensive discovery-and- finding tool as described in focus groups. the result is an application called laneconnex. n design objectives the overall goal of laneconnex is to create a simple, fast search across multiple licensed, open-access, and special-object local knowledge sources that depackages and reaggregates information on the basis of stanford institutional roles. the content of lane’s digital collec- tion includes forty-five hundred journal titles and forty- two thousand other digital resources, including video lectures, executable software, patient handouts, bioin- formatics tools, and a significant store of digitized his- torical materials as a result of the google books program. media types include html pages, pdf documents, jpeg images, mp audio files, mpeg videos, and executable applications. more than three hundred reference titles have been licensed specifically for clinicians at the point of care (e.g., uptodate, emedicine, stat-ref, and micromedex clinical evidence). clinicians wanted their results to reflect subcomponents of a package (e.g., results from the micromedex patient handouts). other clinical content is institutionally managed (e.g., institutional formulary, lab test database, or patient handouts). more than bio- medical research tools have been licensed or selected from open-access content. the needs of biomedical researchers include molecular biology tools and software, biomedi- cal literature databases, citation analysis, chemical and engineering databases, expertise-finding tools, laboratory tools and supplies, institutional-research resources, and upcoming seminars. the specific objectives of the search application are the following: n the user interface should be fast, simple, and intui- tive, with embedded suggestions for improving search results (e.g., did you mean? didn’t find it? have you tried?). n search results from disparate local and external systems should be integrated into a single display based on popular search-engine models familiar to the target population. n the query-retrieval and results display should be separated and reusable to allow customization by role or domain and future expansion into other institutional tools. n resource results should be ranked by relevance and filtered by genre. n metasearch results should be hit counts and fil- tered by category for speed and breadth. results should be reusable for specific views by role. n finding a known article or journal should be streamlined and directly link to the item or “get item” option. n the most popular search options (pubmed, google, and lane journals) should be ubiquitous. n alternative pathways should be dynamic and interactive at the point of need to avoid backtrack- ing and dead ends. n user behavior should be tracked by search term, resource used, and user location to help the library make informed decisions about licensing, meta- data, and missing content. n off-the-shelf software should be used when avail- able or appropriate with development focused on search integration. n the application should be built upon existing metadata-creation systems and trusted web- development technologies. based on these objectives, we designed an application that is an extension of existing systems and technolo- gies. resources are acquired and metadata are provided using the voyager integrated library system (ils). the sfx openurl link resolver provides full-text article access and expands the title search beyond biomedicine to all online journals at stanford. ezproxy provides seamless off-campus access. webtrends provides usage tracking. movable type is used to create faq and help information. a locally developed metasearch application provides a cross search with hit results from more than three hundred external and internal full-text sources. the technologies used to build laneconnex and integrate all of these systems include extensible stylesheet language laneconnex | ketchell et al. transformations (xslt), java, javascript, the apache cocoon project, and oracle. n systems description architecture laneconnex is built on a principle of separation of concerns. the lane content owner can directly change the inclusion of search results, how they are displayed, and additional path-finding information. application programmers use java, javascript, xslt, and structured query language (sql) to create components that generate and modify the search results. the merger of content design and search results occurs “just in time” in the user’s browser. we use component-oriented programming design whereby services provided within the application are defined by simple contracts. in laneconnex, these com- ponents (called “transformers”) consume xml informa- tion and, after transforming it in some way, pass it on to some other component. a particular contract can be fulfilled in different ways for different purposes. this component architecture allows for easy extension of the underlying apache cocoon application. if laneconnex needs to transform some xml data that is not possible with built-in cocoon transformers, it is a simple matter to create a software component that does what is needed and fulfills the transformer contract. apache cocoon is the underlying architecture for laneconnex, as illustrated in figure . this java servlet is an xml–publishing engine that is built upon a compo- nent framework and uses a pipeline-processing model. a declarative language uses pattern matching to associate sets of processing components with particular request urls. content can come from a variety of sources. we use content from the local file system, network file sys- tem, http, and a relational database. the xslt language is used extensively in the pipelines and gives fine control of individual parts of the documents being processed. the end of processing is usually an xhtml document but can be any common mime type. we use cocoon to separate areas of concern so things like content, look and feel, and processing can all be managed as separate entities by different groups of people with little effect on another area. this separation of concerns is manifested by template documents that contain most of the html content common to all pages and are then combined with content documents within a processing pipeline. the declarative nature of the sitemap language and xslt facilitate rapid development with no need to redeploy the entire application to make changes in its behavior. the laneconnex search is composed of several com- ponents integrated into a query-and-results interface: oracle resource metadata, full-text metasearch application, movable type blogging software, “did you mean?” spell checker, ezproxy remote access, and webtrends tracking. n full-text metasearch integration of results from lane’s metasearch applica- tion illustrates cocoon’s many strengths. when a user searches laneconnex, cocoon sends his or her query to the metasearch application, which then dispatches the request to multiple external, full-text search engines and content stores. some examples of these external resources are uptodate, access medicine, micromedex, pubmed, and md consult. the metasearch application interacts with these external resources through jakarta commons http clients. responses from external resources are turned into w c document object model (dom) objects, and xpath expressions are used to resolve hit counts from the dom objects. as result counts are returned, they are added to an xml–based result list and returned to cocoon. the power of cocoon becomes evident as the xml– based metasearch result list is combined with a separate display template. this template-based approach affords content curators the ability to directly add, group, and describe metasearch resources using the language and look that is most meaningful to their specific user communities. for example, there are currently eight metasearch templates curated by an informationist in partnership with a target community. curating these tem- plates requires little to no assistance from programmers. in lane’s interface, a user’s request was sent to the metasearch application, and the application waited five seconds before responding to give external resources a chance to return a result. hit counts in the user interface included a link to refresh and retrieve more results from external resources that had not yet responded. usability studies showed this to be a significant user barrier, since the refresh link was rarely clicked. the initial five second delay also gave users the impression that the site was slow. the laneconnex application makes heavy use of javascript to solve this problem. after a user makes her initial request, javascript is used to poll the metasearch application (through cocoon) on the user’s behalf, pop- ping in result counts as external resources respond. this adds a level of interactivity previously unavailable and makes the metasearch piece of laneconnex much more successful than its previous version. resource metadata laneconnex replaces the catalog as the primary discov- ery interface. metadata describing locally owned and information technology and libraries | march licensed resources (journals, databases, books, videos, images, calculators, and software applications) are stored in the library’s current system of record, an instance of the voyager ils. laneconnex makes no attempt to replace voyager ’s strengths as an application for the selection, acquisition, description, and management of access to library resources. it does, however, replace voyager ’s discovery interface. to this end, metadata for about eight thousand digital resources is extracted from voyager ’s oracle database, converted into marcxml, processed with xslt, and stored in a simple relational database (six tables and twenty-nine attributes) to sup- port fast retrieval speed and tight control over search syntax. this extraction process occurs nightly, with incremental updates every five minutes. the oracle text search engine provides functionality anticipated by our internet-minded users. key features are speed and relevance-ranked results. a highly refined results rank- ing insures that the logical title appears in the first few results. a user ’s query is parsed for wildcard, boolean, proximity, and phrase operators, and then translated into an sql query. results are then transformed into a display version. related services laneconnex compares a user’s query terms against a dictionary. each query is sent to a cocoon spell-checking component that returns suggestions where appropri- ate. this component currently uses the simple object figure . laneconnex architecture. laneconnex | ketchell et al. access protocol (soap)–based spell- ing service from google. google was chosen over the national center for biotechnology information (ncbi) spelling service because of the breadth of terms entered by users; however, cocoon’s component-oriented archi- tecture would make it trivial to change spell checkers in the future. each query is also compared against stanford’s openurl link resolver (findit@stanford). client-side javascript makes a cocoon-mediated query of findit@stanford. using xslt, findit@stanford responses are turned into javascript object notation (json) objects and popped into the interface as appropriate. although the vast majority of laneconnex searches result in zero findit@stanford results, the convenience of searching all of lane’s systems in a single, unified interface far outweighs the effort of implementation. a commercial analytics tool called webtrends is used to collect web statis- tics for making data-centric decisions about interface changes. webtrends uses client-side javascript to track specific user click events. libraries need to track both on-site clicks (e.g., the user clicked on “clinical portal” from the home page) and off-site clicks (e.g., the user clicked on “yamada’s gastroenterology” after doing a search for “ibs”). to facilitate off-site click capture, webtrends requires every external link to include a snippet of javascript. requiring content creators to input this code by hand would be error prone and tedious. laneconnex automatically supplies this code for every class of link (search or static). this specialized webtrends method provides lane with data to inform both interface design and licensing decisions. n results laneconnex version . was released to the stanford biomedical community in july . the current applica- tion can be experienced at http://lane.stanford.edu. the figure . laneconnex resource search results. resource results are ranked by rel- evance. single word titles are given a higher weight in the ranking algorithm to insure they are displayed in the first five results. uniform titles are used to co-locate versions (e.g., the three instances of science from different producers). journals titles are linked to their respective impact factor page in the isi web of knowledge. digital formats that require spe- cial players or restrictions are indicated. the metadata searched for ejournals, databases, ebooks, biotools, video, and medcalcs are lane’s digital resources extracted from the inte- grated library system into a searchable oracle database. the first “all” tab is the combined results of these genres and the lane site help and information. figure . laneconnex related services search enhancements. laneconnex includes a spell checker to avoid a common failure in user searches. ajax services allow the inclusion of search results from other sources for common zero results failures. for example, the stanford link resolver database is simultaneously searched to insure online journals outside the scope of biomedicine are presented as a linked result for the user. production version has proven reliable over two years. incremental user focus groups have been employed to improve the interface as issues arose. a series of vignettes will be used to illustrate how the current version of information technology and libraries | march the “sunetid login” is required. n user query: “new yokrer.” a faculty member is looking for an arti- cle in the new yorker for a class reading assignment. he makes a typing error, which invokes the “did you mean?” function (see figure ). he clicks on the correct spelling. no results are found in the resource search, but a simul- taneous search of the link-resolver database finds an instance of this title licensed for the campus and displays a clickable link for the user. n user query: “pathway analy- sis.” a post–doc is looking for infor- mation on how to share an ingenuity pathway. figure illustrates the inte- gration of the locally created lane faqs. faqs comprise a broad spec- trum of help and how-to information as described by our focus groups. help text is created in the movable type blog software, and made searchable through the laneconnex application. the movable type interface lowers the barrier to html content creation by any staff member. more complex answers include embedded images and videos to enable the user to see exactly how to do a particular proce- dure. cocoon allows for the syndica- tion of subsets of this faq content back into static html pages where it can be displayed as both category-specific lists or as the text for scroll-over help for a link. having a single store of help information insures the content is updated once for all instances. n user query: “uterine cancer kapp.” a resident is looking for a known article. laneconnex simultaneously searches pubmed to increase the likelihood of user success (see figure ). clicking on the pubmed tab retrieves the results in the native interface; however, the user sees the pubmed@stanford ver- sion, which includes embedded links to the article based on our openurl link resolver. the ability to retrieve results from bibliographic databases that includes article resolution insures that our biomedical community is always using the correct url to insure maximum full-text article access. user testing in found that adding the three most frequently used sources (pubmed, google, and lane catalog) into our one-box laneconnex search was a significant time saver. it addresses laneconnex meets the design objectives from the user’s perspective. n user query: “science.” a graduate student is look- ing for the journal science. the laneconnex results are listed in relevance order (see figure ). single- word titles are given a higher weight in the rank- ing algorithm to insure they are displayed in the first five results. results from local metadata are displayed by uniform title. for example, lane has three instances of the journal science, and each version is linked to the appropriate external store. brief notes provide critical information for particu- lar resources. for example, restricted local patient education documents and video seminars note that figure . example of integration of local content stores. help information is managed in moveable type and integrated into laneconnex search results. laneconnex | ketchell et al. the expectation on the part of our users that they could search for an article or a journal title in a single search box without first selecting a database. n user query: “serotonin pul- monary hypertension.” a medical student is looking for the correlation of two topics. clicking on the “clinical” tab, the student sees the results of the clinical metasearch in fig- ure . metasearch results are deep searches of sources within licensed packages (e.g., text- books in md consult or a spe- cific database in micromedex), local content (e.g., stanford’s lab-test database), and open- access content (e.g., ncbi databases). pubmed results are tailored strategies tiered by evidence. for example, the evidence-summaries strategy retrieves results from twelve clinical-evidence resources (e.g., buj, clinical evidence, and cochrane systematic reviews) that link to the full-text licensed by stanford. an example of the bioresearch metasearch is shown in figure . content selected for this audience includes literature databases, funding sources, patents, structures, clinical trials, protocols, and stanford expertise integrated with gene, protein, and phe- notype tools. user testing revealed that many users did not click on the “clinical” tab. the clinical metasearch was originally developed for the clinical portal page and focused on clinicians in practice; however, the results needed to be exposed more directly as part of the laneconnex search. figure illustrates the “have you tried?” feature that displays a few relevant clinical-content sources without requiring the user to select the “clinical” tab. this fea- ture is managed by the smartsearch component of the laneconnex system. smartsearch sends the user’s query terms to pubmed, extracts a subset of articles associated with those terms, extracts the mesh headings for those articles, and computes the frequency of headings in the articles to determine the most likely mesh terms associ- ated with the user’s query terms. these mesh terms are mapped to mesh terms associated with each metasearch resource. preliminary evaluation indicates that the clini- cal content is now being discovered by more users. figure . example of integration of popular search engines into laneconnex results. three of the most popular searches based on usage analysis are included at the top level. pubmed and google are mapped to lane’s link resolver to retrieve the full article. creating or editing metasearch templates is a curator- driven task. programming is only required to add new sources to the metasearch engine. a curator may choose from more than three hundred sources to create a dis- cipline-based layout using general templates. names, categories, and other description information are all at the curator ’s discretion. while developing new sub- specialty templates, we discovered that clinicians were confused by the difference in layout of their specialty portal and their metasearch results (e.g., the cardiology portal used the generic clinical metasearch). to address this issue, we devised an approach that merges a portal and metasearch into a single entity as illustrated in figure . a combination of the component-oriented architecture of laneconnex and javascript makes the integration of metasearch results into a new template patterned after a portal easy to implement. this strategy will enable the creation of templates contextually appropriate to knowl- edge requests originating from electronic medical-record systems in the future. direct user feedback and usage statistics confirm that search is now the dominant mode of navigation. the amount of time each user spends on the website has dropped since the release of version . . we speculate that the integrated search helps our users find relevant information technology and libraries | march information more efficiently. focus groups with students are uniformly positive. graduate students like the ability to find digital articles using a single search box. medical students like the clinical metasearch as an easy way to look up new topics in texts and customized pubmed searches. bioengineering students like the ability to easily look up patient care–related topics. pediatrics residents and attend- ings have championed the develop- ment of their portal and metasearch focused on their patient population. medical educators have commented on their ability to focus on the best information sources. n discussion a review of websites in found that most biomedical libraries had sep- arate search interfaces for their digital resources, library catalog, and exter- nal databases. biomedical libraries are implementing metasearch software to cross search proprietary data- bases. the university of california, davis is using the metalib software to federate searching multiple bib- liographic databases. the university of south california and florida state university are using webfeat soft- ware to search clinical textbooks. the health sciences library system at the university of pittsburgh is using vivisimo to search clinical textbooks and bioresearch tools. academic libraries are introducing new “resource shopping” applications, such as the endeca project at north carolina state university, the summa project at the university of aarhus, and the vufind project at villanova university. these systems offer a single query box, faceted results, spell checking, recom- mendations based on user input, and asynchronous javascript and xml (ajax) for live status information. we believe our approach is a practi- cal integration for our biomedical com- munity that bridges finding a resource and finding a specific item through figure . integration of metasearch results into laneconnex. results from two general, role-based metasearches (bioresearch and clinical) are included in the laneconnex interface. the first image shows a clinician searching laneconnex for serotonin pulmonary hypertension. selecting the clinical tab presents the clinical content metasearch display (second image), and is placed deep inside the source by selecting a title (third image). laneconnex | ketchell et al. a metasearch of multiple databases. the laneconnex application searches across digital resources and external data stores simultaneously and pres- ents results in a unified display. the limitation to our approach is that the metasearch returns only hit counts rather than previews of the specific content. standardization of results from external systems, particularly receipt of xml results, remains a chal- lenge. federated search engines do integrate at this level, but are usually slow or limit the number of results. true integration awaits health level seven (hl ) clinical decision support standards and national information standards organization (niso) metasearch initiative for query and retrieval of specific content. one of the primary objectives of laneconnex is speed and ease of use. ranking and categorization of results has been very successful in the eyes of the user community. the integration of metasearch results has been par- ticularly successful with our pediatric specialty portal and search. however, general user understanding of how the clinical and biomedical tabs related to the genre tabs in laneconnex has been problematic. we reviewed web engines and found a similar challenge in presenting disparate format results (e.g., video or image search results) or lists of hits from different systems (e.g., ncbi’s entrez search results). we are continuing to develop our new specialty portal-and-search model and our smartsearch term-mapping com- ponent to further integrate results. n conclusion laneconnex is an effective and open- ended search infrastructure for inte- grating local resource metadata and full-text content used by clinicians and biomedical researchers. its effective- ness comes from the recognition that users prefer a single query box with relevance or categorically organized results that lead them to the most likely figure . example of a bioresearch metasearch. figure . the smartsearch component embeds a set of the metasearch results into the laneconnex interface as “have you tried?” clickable links. these links are the equivalent of selecting the title from a clinical metasearch result. the example search for atypical malig- nant rhabdoid tumor (a rare childhood cancer) invokes oncology and pediatric textbook results. these texts and pubmed provide quick access for a medical student or resident on the pediatric ward. figure . example of a clinical specialty portal with integrated metasearch. clinical portal pages are organized so metasearch hit counts can display next to content links if a user executes a search. this approach removes the dissonance clinicians felt existed between separate portal page and metasearch results in version . . information technology and libraries | march answer to a question or prospects in their exploration. the application is based on separation of concerns and is easily extensible. new resources are constantly emerg- ing, and it is important that libraries take full advantage of existing and forthcoming content that is tailored to their user population regardless of the source. the next major step in the ongoing development of laneconnex is becoming an invisible backend application to bring content directly into the user’s workflow. n acknowledgements the authors would like to acknowledge the contribu- tions of the entire laneconnex technical team, in par- ticular pam murnane, olya gary, dick miller, rick zwies, and rikke ogawa for their design contributions, philip constantinou for his architecture contribution, and alain boussard for his systems development contributions. references . denise t. covey, “the need to improve remote access to online library resources: filling the gap between com- mercial vendor and academic user practice,” portal libraries and the academy no. ( ): – ; nobert lossau, “search engine technology and digital libraries,” d-lib magazine no. ( ), www.dlib.org/dlib/june /lossau/ lossau.html (accessed mar. , ); oclc, “college students’ perception of libraries and information resource,” www.oclc.org/reports/ perceptionscollege.htm (accessed mar , ); and jim hender- son, “google scholar: a source for clinicians,” canadian medical association journal no. ( ). . covey, “the need to improve remote access to online library resources”; lossau, “search engine technology and digital libraries”; oclc, “college students’ perception of libraries and information resource.” . jane lee, “uc health sciences metasearch exploration. part : graduate student gocus group findings,” uc health sciences metasearch team, www.cdlib.org/inside/assess/ evaluation_activities/docs/ /draft_gradreport_march . pdf (accessed mar. , ). . karen k. grandage, david c. slawson, and allen f. shaughnessy, “when less is more: a practical approach to searching for evidence-based answers,” journal of the medical library association no. ( ): – . . nicola cannata, emanuela merelli, and russ b. altman, “time to organize the bioinformatics resourceome,” plos com- putational biology no. ( ): e . . craig silverstein et al., “analysis of a very large web search engine query log,” www.cs.ucsb.edu/~almeroth/ classes/tech-soc/ -winter/papers/analysis.pdf (accessed mar. , ); anne aula, “query formulation in web informa- tion search,” www.cs.uta.fi/~aula/questionnaire.pdf (accessed mar. , ); jorge r. herskovic, len y. tanaka, william hersh, and elmer v. bernstam, “a day in the life of pubmed: analysis of a typical day’s query log,” journal of the american medical informatics association no. ( ): – . . herskovic, “a day in the life of pubmed.” . davis libraries university of california, “quicksearch,” http://mysearchspace.lib.ucdavis.edu/ (accessed mar. , ). . eileen eandi, “health sciences multi-ebook search,” norris medical library newsletter (spring ), norris medical library, university of southern california, www.usc.edu/hsc/ nml/lib-information/newsletters.html (accessed mar. , ); maguire medical library, florida state university, “webfeat clinical book search,” http://med.fsu.edu/library/tutorials/ webfeat _viewlet_swf.html (accessed mar. , ). . jill e. foust, philip bergen, gretchen l. maxeiner, and peter n. pawlowski, “improving e-book access via a library- developed full-text search tool,” journal of the medical library association no. ( ): – . . north carolina state university libraries, “endeca at the ncsu libraries,” www.lib.ncsu.edu/endeca (accessed mar. , ); hans lund, hans lauridsen, and jens hofman han- sen, “summa—integrated search,” www.statsbiblioteket.dk/ publ/summaenglish.pdf (accessed mar. , ); falvey memo- rial library, villanova university, “vufind,” www.vufind.org (accessed mar. , ). . see the health level seven (hl ) clinical decision sup- port working committee activities, in particular the infobutton standard proposal at www.hl .org/special/committees/dss/ index.cfm and the niso metasearch initiative documentation at www.niso.org/workrooms/mi (accessed mar , ). . national center for biotechnology information (ncbi) entrez cross-database search, www.ncbi.nlm.nih.gov/entrez (accessed mar. , ). acrl alcts lita cover , cover jaunter cover index to advertisers information technology and libraries | march tagging: an organization scheme for the internet marijke a. visser how should the information on the internet be organized? this question and the possible solutions spark debates among people concerned with how we identify, classify, and retrieve internet content. this paper discusses the benefits and the controversies of using a tagging system to organize internet resources. tagging refers to a clas- sification system where individual internet users apply labels, or tags, to digital resources. tagging increased in popularity with the advent of web . applications that encourage interaction among users. as more information is available digitally, the challenge to find an organiza- tional system scalable to the internet will continue to require forward thinking. trained to ensure access to a range of informational resources, librarians need to be concerned with access to internet content. librarians can play a pivotal role by advocating for a system that sup- ports the user at the moment of need. tagging may just be the necessary system. w ho will organize the information available on the internet? how will it be organized? does it need an organizational scheme at all? in , thomas and griffin asked a similar question, “who will create the metadata for the internet?” in their article with the same name. ten years later, this question has grown beyond simply supplying metadata to assuring that at the moment of need, someone can retrieve the information necessary to answer their query. given new classification tools available on the internet, the time is right to reas- sess traditional models, such as controlled vocabularies and taxonomies, and contrast them with folksonomies to understand which approach is best suited for the future. this paper gives particular attention to delicious, a social networking tool for generating folksonomies. the amount of information available to anyone with an internet connection has increased in part because of the internet’s participatory nature. users add content in a variety of formats and through a variety of applications to personalize their web experience, thus making internet content transitory in nature and challenging to lock into place. the continual influx of new information is caus- ing a rapid cultural shift, more rapid than many people are able to keep up with or anticipate. conversations on a range of topics that take place using web technologies happen in real time. unless you are a participant in these conversations and debates using web-based communica- tion tools, changes are passing you by. internet users in general have barely grasped the concept of web . and already the advanced “internet cognoscenti” write about web . . regarding the organization and availability of internet content, librarians need to be ahead of the crowd as the voice who will assure content will be readily accessible to those that seek it. internet users actively participat- ing in and shaping the online communities are, perhaps unintentionally, influencing how those who access infor- mation via the internet expect to be able to receive and use digital resources. librarians understand that the way information is organized is critical to its accessibility. they also understand the communities in which they operate. today, librarians need to be able to work seam- lessly among the online communities, the resources they create, and the end user. as internet use evolves, librar- ians as information stakeholders should stay abreast of web . developments. by positioning themselves to lead the future of information organization, librarians will be able to select the best emerging web-based tools and applications, become familiar with their strengths, and leverage their usefulness to guide users in organizing internet content. shirky argues that the internet has allowed new com- munities to form. primarily online, these communities of internet users are capable of dramatically changing society both on- and offline. shirky contends that because of the internet, “group action just got easier.” according to shirky, we are now at the critical point where internet use, while dependent on technology, is actually no longer about the technology at all. the web today (web . ) is about participation. “this [the internet] is a medium that is going to change society.” lessig points out that content creators are “writing in the socially, culturally relevant sense for the st century and to be able to engage in this writing is a measure of your literacy in the st century.” it is significant that creating content is no longer reserved for the internet cognoscenti. internet users with a variety of technological skills are participating in web . com- munities. information architects, web designers, librarians, busi- ness representatives, and any stakeholder dependent on accessing resources on the internet have a vested interest in how internet information is organized. not only does the architecture of participation inherent in the internet encourage completely new creative endeavors, it serves as a platform for individual voices as demonstrated in marijke a. visser (marijkea@gmail.com) is a library and infor- mation science graduate student at indiana university, india- napolis, and will be graduating may . she is currently work- ing for ala’s office for information and technology policy as an information technology policy analyst, where her area of focus includes telecommunications policy and how it affects access to information. tagging: an organization scheme for the internet | visser personal and organizationally sponsored blogs: lessig . , boing boing, open access news, and others. these internet conversations contribute diverse viewpoints on a stage where, theoretically, anyone can access them. web . technologies challenge our understanding of what con- stitutes information and push policy makers to negotiate equitable internet-use policies for the public, the content creators, corporate interests, and the service providers. to maintain an open internet that serves the needs of all the players, those involved must embrace the opportunity for cultural growth the social web represents. for users who access, create, and distribute digital content, information is anything but static; nor is using it the solitary endeavor of reading a book. its digital format makes it especially easy for people to manipulate it and shape it to create new works. people are sharing these new works via social technologies for others to then remix into yet more distinct creative work. communication is fundamentally altered by the ability to share content on the internet. today’s internet requires a reevaluation of how we define and organize information. the manner in which digital information is classified directly affects each user’s ability to access needed information to fully participate in twenty-first-century culture. new para- digms for talking about and classifying information that reflect the participatory internet are essential. n background the controversy over organizing web-based information can be summed up comparing two perspectives repre- sented by shirky and peterson. both authors address how information on the web can be most effectively orga- nized. in her introduction, peterson states, “items that are different or strange can become a barrier to networking.” shirky maintains, “as the web has shown us, you can extract a surprising amount of value from big messy data sets.” briefly, in this instance ontology refers to the idea of defining where digital information can and should be located (virtually). folksonomy describes an organiza- tional system where individuals determine the placement and categorization of digital information. both terms are discussed in detail below. although any organizational system necessitates talking about the relationship(s) among the materials being organized, the relationships can be classified in multiple ways. to organize a given set of entities, it is necessary to establish in what general domain they belong and in what ways they are related. applying an ontological, or hierar- chical, classification system to digital information raises several points to consider. first, there are no physical space restrictions on the internet, so relationships among digital resources do not need to be strictly identified. second, after recognizing that internet resources do not need the same classification standards as print material, librarians can begin to isolate the strengths of current nondigital systems that could be adapted to a system for the internet. third, librarians must be ready to eliminate current systems entirely if they fail to serve the needs of internet users. traditional systems for organizing information were developed prior to the information explosion on the internet. the internet’s unique platform for creating, storing, and disseminating information challenges pre– digital-age models. designing an organizational system for the internet that supports creative innovation and succeeds in providing access to the innovative work is paramount to moving the twenty-first-century culture forward. n assessing alternative models controversy encourages scrutiny of alternative models. in understanding the options for organizing digital infor- mation, it is important to understand traditional classifi- cation models. smith discusses controlled vocabularies, taxonomies, and facets as three traditional methods for applying metadata to a resource. according to smith, a controlled vocabulary is an unambiguous system for managing the meanings of words. it links synonyms, allowing a search to retrieve information on the basis of the relationship between synonyms. taxonomies are hierarchical, controlled vocabularies that establish par- ent–child relationships between terms. a faceted classifi- cation system categorizes information using the distinct properties of that information. in such a system, infor- mation can exist in more than one place at a time. a fac- eted classification system is a precursor to the bottom-up system represented by folksonomic tagging. folksonomy, a term coined in by thomas vander wal, refers to a “user-created categorical structure development with an emergent thesaurus.” vander wal further separates the definition into two types: a narrow and a broad folk- sonomy. in a broad folksonomy, many people tag the same object with numerous tags or a combination of their own and others’ tags. in a narrow folksonomy, one or few people tag an object with primarily singular terms. internet searching represents a unique challenge to people wanting to organize its available information. search engines like yahoo! and google approach the cha- otic mass of information using two different techniques. yahoo! created a directory similar to the file folder system with a set of predetermined categories that were intended to be universally useful. in so doing, the yahoo! devel- opers made assumptions about how the general public would categorize and access information. the categories information technology and libraries | march and subsequent subcategories were not necessarily logi- cally linked in the eyes of the general public. the yahoo! directory expanded as internet content grew, but the digi- tal folder system, like a taxonomy, required an expert to maintain. shirky notes the yahoo! model could not scale to the internet. there are too many possible links to be able to successfully stay within the confines of a hierar- chical classification system. additionally, on the internet, the links are sufficient for access because if two items are linked at least once, the user has an entry point to retrieve either one or both items. a hierarchical system does not assure a successful internet search and it requires a user to comprehend the links determined by the managing expert. in the google approach, developers acknowl- edged that the user with the query best understood the unique reasoning behind her search. the user therefore could best evaluate the information retrieved. according to shirky, the google model let go of the hierarchical file system because developers recognized effective search- ing cannot predetermine what the user wants. unlike yahoo!, google makes the links between the query and the resources after the user types in the search terms. trusting in the link system led google to understand and profit from letting the user filter the search results. to select the best organizational model for the internet it is critical to understand its emergent nature. a model that does not address the effects of web . on internet use and fails to capture participant-created content and tagging will not be successful. one approach to orga- nizing digital resources has been for users to bookmark websites of personal interest. these bookmarks have been stored on the user’s computer, but newer models now combine the participatory web with saving, or tagging, websites. social bookmarking typifies the emergent web and the attraction of online networking. innovative and controversial, the folksonomy model brings to light numerous criteria necessary for a robust organizational system. a social bookmarking network, delicious is a tool for generating folksonomies. it com- bines a large amount of self-interest with the potential for an equal, if not greater, amount of social value. delicious users add metadata to resources on the internet by apply- ing terms, or tags, to urls. users save these tagged web- sites to a personal library hosted on the delicious website. the default settings on delicious share a user’s library publicly, thus allowing other people—not limited to reg- istered delicious account holders—to view any library. that the delicious developers understood how internet users would react to this type of interactive application is reflected in the popularity of delicious. delicious arrived on the scene in , and in developers introduced a number of features to encourage further user collabora- tion. with a new look (going from the original del.icio.us to its current moniker, delicious) as well as more ways for users to retrieve and share resources by , delicious had million registered users and million unique urls. the reputation of delicious has generated inter- est among people concerned with organizing the infor- mation available via the internet. how does the folksonomy or delicious model of open-ended tagging affect searching, information retriev- ing, and resource sharing? delicious, whose platform is heavily influenced by its users, operates with no hier- archical control over the vocabulary used as tags. this underscores the organization controversy. bottom-up tagging gives each person tagging an equal voice in the categorization scheme that develops through the user generated tags. at the same time, it creates a chaotic infor- mation-retrieval system when compared to traditional controlled vocabularies, taxonomies, and other methods of applying metadata. a folksonomy follows no hier- archical scheme. every tag generated supplies personal meaning to the associated url and is equally weighted. there will be overlap in some of the tags users select, and that will be the point of access for different users. for the unique tags, each delicious user can choose to adopt or reject them for their personal tagging system. either way, the additional tags add possible future access points for the rest of the user community. the social usefulness of the tags grows organically in relationship to their adop- tion by the group. can the internet support an organizational system controlled by user-generated tags? by the very nature of the participatory web, whose applications often get bet- ter with user input, the answer is yes. delicious and other social tagging systems are proving that their folksonomic approach is robust enough to satisfy the organizational needs of their users. defined by vander wal, a broad folk- sonomy is a classification system scalable to the internet. the problem with projecting already-existing search and classification strategies to the internet is that the internet is constantly evolving, and classic models are quickly overcome. even in the nonprint world of the internet, taxonomies and controlled vocabulary entail a commitment both from the entity wanting to organize the system and the users who will be accessing it. developing a taxonomy involves an expert, which requires an outlay of capital and, as in the case with yahoo!, a taxonomy is not necessarily what users are looking for. to be used effectively, taxonomies demand a certain amount of user finesse and complacency. the user must understand the general hierarchy and by default must suspend their own sense of category and subcategory if they do not mesh with the given system. the search model used by google, where the user does the filtering, has been a significantly more successful search engine. google recognizes natural language, making it user friendly; however, it remains merely a search engine. it is successful at making links, but it leaves the user stranded without a means to orga- nize search results beyond simple page rank. traditional tagging: an organization scheme for the internet | visser hierarchical systems and search strategies like those of yahoo! and google neglect to take into account the tre- mendous popularity of the participatory web. successful web applications today support user interaction; to disre- gard this is naive and short-sighted. in contrast to a simple page-rank results list or a hierarchical system, delicious results provide the user with rich, multilayer results. figure shows four of the first ten results of a delicious search for the term “folk- sonomy.” the articles by the four authors in the left col- umn were tagged according to the diagram. two of the articles are peer-reviewed, and two are cited repeatedly by scholars researching tagging and the internet. in this example, three unique terms are used to tag those articles, and the other terms provide additional entry points for retrieval. further information available using delicious shows that the guy article was tagged by , users, the mathes article by , users, the shirky article by , users, and the peterson article by users. from the basic delicious search, the user can combine terms to narrow the query as well as search what other users have tagged with those terms. similar to the card catalog, where a library patron would often unintentionally find a book title by browsing cards before or after the actual title she originally wanted, a delicious user can browse other users’ libraries, often finding additional pertinent resources. a user will return a greater number of relevant and automatically filtered results than with an advanced google search. as an ancillary feature, once a delicious user finds an attractive tag stream—a series of tags by a particular user—they can opt to follow the user who created the tag stream, thereby increasing their personal resources. hence delicious is effective personally and socially. it emulates what internet users expect to be able to do with digital content: find interesting resources, per- sonalize them, in this case with tags, and put them back out for others to use if they so choose. proponents of folksonomy recognize there are ben- efits to traditional taxonomies and controlled vocabulary systems. shirky delineates two features of an organi- zational system and their characteristics, providing an example of when a hierarchical system can be successful (see table ). these characteristics apply to situations using data- bases, journal articles, and dissertations as spelled out by peterson, for example. specific organizations with identifiable common terminology—for example, medical libraries—can also benefit from a traditional classification system. these domains are the antithesis of the domain represented by the web. the success of controlled vocab- ularies, taxonomies, and their resulting systems depends on broad user adoption. that, in combination with the cost of creating and implementing a controlled system, raises questions as to their utility and long-term viability for use on the web. though meant for longevity, a taxonomy fulfills a need at one fixed moment in time. a folksonomy is never static. taxonomies developed by experts have not yet been able to be extended adequately for the breadth and depth of internet resources. neither have traditional viewpoints been scaled to accept the challenges encountered in try- ing to organize the internet. folksonomy, like taxonomy, seeks to provide the information critical to the user at the moment of need. folksonomy, however, relies on users to create the links that will retrieve the desired results. doctorow puts forward three critiques of a hierarchical metadata system, emphasizing the inadequacies of apply- ing traditional classification schemes to the digital stage: . there is not a “correct” way to categorize an idea. . competing interests cannot come to a consensus figure . search results for “folksonomy” using delicious. table . domains and their participants domain to be organized participants in the domain small corpus expert catalogers formal categories authoritative source of judgment restricted entities coordinated users clear edges expert users information technology and libraries | march on a hierarchical vocabulary. . there is more than one way to describe some- thing. doctorow elaborates: “requiring everyone to use the same vocabulary to describe their material denudes the cognitive landscape, enforces homogeneity in ideas.” the internet raises the level of participation to include innumerable voices. the astonishing thing is that it thrives on this participation. guy and tonkin address the “folksonomic flaw” by saying user-generated tags are by definition imprecise. they can be ambiguous, overly personal, misspelled, and a contrived compound word. guy and tonkin suggest the need to improve tagging by educating the users or by improving the systems to encourage more accurate tagging. this, however, does not acknowledge that successful web . applications depend on the emergent wisdom of the user community. the systems permit organic evolution and continual improvement by user participation. a folksonomy evolves much the way a spe- cies does. unique or single-use tags have minimal social import and do not gain recognition. tags used by more than a few people reinforce their value and emerge as the more robust species. n conclusion the benefits of the internet are accessible to a wide range of users. the rewards of participation are imme- diate, social, and exponential in scope. user-generated content and associated organization models support the internet’s unique ability to bring together unlikely social relationships that would not necessarily happen in another milieu. to paraphrase shirky and lessig, people are participating in a moment of social and technologi- cal evolution that is altering traditional ways of thinking about information, thereby creating a break from tradi- tional systems. folksonomic classification is part of that break. its utility grows organically as users add tagged content to the system. it is adaptive, and its strengths can be leveraged according to the needs of the group. while there are “folksonomic flaws” inherent in a bottom- up classification system, there is tremendous value in weighting individual voices equally. following the logic of web . technology, folksonomy will improve accord- ing to the input of the users. it is an organizational system that reflects the basic tenets of the emergent internet. it may be the only practical solution in a world of participa- tory content creation. shirky describes the internet by saying, “there is no shelf in the digital world.” classic organizational schemes like the dewey decimal system were created to organize resources prior to the advent of the internet. a hierarchical system was necessary because there was a physical limita- tion on where a resource could be located; a book can only exist in one place at one time. in the digital world, the shelf is simply not there. material can exist in many different places at once and can be retrieved through many avenues. a broad folksonomy supports a vibrant search strategy. it combines individual user input with that of the group. this relationship creates data sets inherently meaningful to the community of users seeking information on any given topic at any given moment. this is why a folksonomic approach to organizing information on the internet is suc- cessful. users are rewarded for their participation, and the system improves because of it. folksonomy mirrors and supports the evolution of the internet. librarians, trained to be impartial and ethically bound to assure access to information, are the logical mediators among content creators, the architecture of the web, corporate interests, and policy makers. critical con- versations are no longer happening only in traditional publications of the print world. they are happening with communication platforms like youtube, twitter, digg, and delicious. information organization is one issue on which librarians can be progressive. dedicated to making information available, librarians are in a unique position to take on challenges raised by the internet. as the profession experiments with the introduction of web . , librarians need to position themselves between what is known and what has yet to evolve. librarians have always leveraged the interests and needs of their users to tailor their services to the individual entry point of every person who enters the library. because more and more resources are accessed via the internet, librarians will have to maintain a presence throughout the web if they are to continue to speak for the informational needs of their users. part of that presence necessitates an ability to adapt current models to the internet. more importantly, it requires recognition of when to forgo con- ventional service methods in favor of more innovative approaches. working in concert with the early adopters, corporate interests, and general internet users, librarians can promote a successful system for organizing internet resources. for the internet, folksonomic tagging is one solution that will assure users can retrieve information necessary to answer their queries. references and notes . charles f. thomas and linda s. griffin, “who will cre- ate the metadata for the internet?” first monday , no. (dec. ). . web . is a fairly recent term, although now ubiquitous among people working in and around internet technologies. attributed to a conference held in between medialive tagging: an organization scheme for the internet | visser international and o’reilly media, web . refers to the web as being a platform for harnessing the collective power of internet users interested in creating and sharing ideas and information without mediation from corporate, government, or other hierar- chical policy influencers or regulators. web . is a much more fluid concept as of this writing. there are individuals who use it to refer to a semantic web where information is analyzed or processed by software designed specifically for computers to carry out the currently human-mediated activity of assigning meaning to information on a webpage. there are librarians involved with exploring virtual-world librarianship who refer to the d environment as web . . the important point here is that what internet users now know as web . is in the process of being altered by individuals continually experimenting with and improving upon existing web applications. web . is the undefined future of the participatory internet. . clay shirky, “here comes everybody: the power of organizing without organizations” (presentation videocast, berkman center for internet & society, harvard university, cambridge, mass., ), http://cyber.law.harvard.edu/inter active/events/ / /shirky (accessed oct. , ). . ibid. . lawerence lessig, “early creative commons history, my version,” videocast, aug. , , lessig . , http://lessig.org/ blog/ / /early_creative_commons_history.html (accessed aug. , ). . elaine peterson, “beneath the metadata: some philosophi- cal problems with folksonomy,” d-lib magazine , no. ( ), http://www.dlib.org/dlib/november /peterson/ peterson .html (accessed sept. , ). . clay shirky, “ontology is overrated: categories, links, and tags” online posting, spring , clay shirky’s writings about the internet, http://www.shirky.com/writings/ontology_ overrated.html#mind_reading (accessed sept. , ). . gene smith, tagging: people-powered metadata for the social web (berkeley, calif.: new riders, ): . . ibid., . . thomas vander wal, “folksonomy,” online posting, feb. , , vanderwal.net, http://www.vanderwal.net/folksonomy .html (accessed aug. , ). . thomas vander wal, “explaining and showing broad and narrow folksonomies,” online posting, feb. , , personal infocloud, http://www.personalinfocloud.com/ / / explaining_and_.html (accessed aug. , ). . shirky, “ontology is overrated.” . ibid. . michael arrington, “exclusive: screen shots and feature overview of delicious . preview,” online posting, june , , techcrunch, http://www.techcrunch.com/ / / / exclusive-screen-shots-and-feature-overview-of-delicious- -preview/(accessed jan. , ). . smith, tagging, – . . vander wal, “explaining and showing broad and narrow folksonomies.” . adam mathes, “folksonomies—cooperative classifica- tion and communication through shared metadata” (graduate paper, university of illinois urbana–champaign, dec. ); peterson, “beneath the metadata”; shirky, “ontology is over- rated”; thomas and griffin, “who will create the metadata for the internet?” . shirky, “ontology is overrated.” . peterson, “beneath the metadata.” . cory doctorow, “metacrap: putting the torch to seven straw-men of the meta-utopia,” online posting, aug. , , the well, http://www.well.com/~doctorow/metacrap.htm (accessed sept. , ). . marieke guy and emma tonkin, “folksonomies: tidy- ing up tags?” d-lib magazine , no. ( ), http://www.dlib .org/dlib/january /guy/ guy.html (accessed sept. , ). . shirky, “ontology is overrated.” global interoperability continued from page . julie renee moore, “rda: new cataloging rules, com- ing soon to a library near you!” library hi tech news , no. , ( ): . . rick bennett, brian f. lavoie, and edward t. o’neill, “the concept of a work in worldcat: an application of frbr,” library collections, acquisitions, & technical services , no. , ( ): . . park, “cross-lingual name and subject access.” . ibid. . thomas b. hickey, “virtual international authority file” (microsoft powerpoint presentation, ala annual conference, new orleans, june ), http://www.oclc.org/research/ projects/viaf/ala c.ppt (accessed dec. , ). . leaf, “leaf project consortium,” http://www.crxnet .com/leaf/index.html (accessed dec. , ). . bennett, lavoie, and o’neill, “the concept of a work in worldcat.” . alan danskin, “mature consideration: developing biblio- graphic standards and maintaining values,” new library world , no. / , ( ): . . ibid. . bennett, lavoie, and o’neill, “the concept of a work in worldcat.” . moore, “rda.” . danskin, “mature consideration,” . . ibid.; park, “cross-lingual name and subject access.” information technology and libraries | march andrew k. pace president’s message: lita now andrew k. pace (pacea@oclc.org) is lita president / and executive director, networked library services at oclc inc. in dublin, ohio. a t the time of this writing, my term as lita presi- dent is half over; by the time of publication, i will be in the home stretch—a phrase that, to me, always connotes relief and satisfaction that is never truly realized. i hope that this time between ala conferences is a time of reflection for the lita board, committees, inter- est groups, and the membership at large. various strate- gic planning sessions are, i hope, leading us down a path of renewal and regeneration of the division. of course, the world around us will have its effect—in particular, a political and economic effect. first, the politics. i was asked recently to give my opinion about where the new administration should focus its attention regarding library technology. i had very little time to think of a pithy answer to this ques- tion, so i answered with my gut that the united states needs to continue its investment in it infrastructure so that we are on par with other industrialized nations while also lending its aid to countries that are lagging behind. furthermore, i thought it an apt time to redress issues of data privacy and retention. the latter is often far from our minds in a world more connected, increasingly through wireless technology, and with a user base that, as one privacy expert put it, would happily trade a dna sample for an extra value meal. i will resist the urge to write at greater length a treatise on the bill of rights and its status in . i will hope, however, that lita’s technology and access and legislation and regulation committees will feel reinvigorated post–election and post–inauguration to look carefully at the issues of it policy. our penchant for new tools should always be guided and tempered by the implementation and support of policies that rational- ize their use. as for the economy, it is our new backdrop. one anecdotal view of this is the number of e-mails i’ve received from committee appointees apologizing that they will not be able to attend ala conferences as planned because of the economic downturn and local cuts to library budgets. libraries themselves are in a paradoxical situation—increasing demand for the free services that libraries offer while simultaneously facing massive budget cuts that support the very collections and programs people are demanding. what can we do? well, i would suggest that we look at library technology through a lens of efficiency and cost savings, not just from a perspective of what is cool or trendy. when it comes to running systems, we need to keep our focus on end-user satisfaction while consider- ing total cost of ownership. and if i may be selfish for a moment, i hope that we will not abandon our profes- sional networks and volunteer activities. while we all make sacrifices of time, money, and talent to support our profession, it is often tempting when economic times are hard to isolate ourselves from the professional networks that sustain us in times of plenty. politics and economics? though i often enjoy being cynical, i also try to make lemonade from lemons when- ever i can. i think there are opportunities for libraries to get their own economic bailout in supporting public works and emphasizing our role in contributing to the public good. we should turn our “woe-are-we” tenden- cies that decry budget cuts and low salaries into champi- oned stories of “what libraries have done for you lately.” and we should go back to the roots of it, no matter how mythical or anachronistic, and think about what we can do technically to improve systemwide efficiencies. i encourage the membership to stay involved and reengage, whether through direct participation in lita activities or through a closer following of the activities in the ala office of information technology policy (oitp, www.ala.org/ala/aboutala/offices/oitp) and the ala washington office itself. there is much to follow in the world that affects our profession, and so many are doing the heavy lifting for us. all we need to do sometimes is pay attention. make fun of me if you want for stealing a campaign phrase from richard nixon, but i kept coming back to it in my head. in short, library information technology— now more than ever. information technology and libraries | march t. michael silver monitoring network and service availability with open-source software silver describes the implementation of a monitoring sys- tem using an open-source software package to improve the availability of services and reduce the response time when troubles occur. he provides a brief overview of the litera- ture available on monitoring library systems, and then describes the implementation of nagios, an open-source network monitoring system, to monitor a regional library system’s servers and wide area network. particular atten- tion is paid to using the plug-in architecture to monitor library services effectively. the author includes example displays and configuration files. editor’s note: this article is the winner of the lita/ex libris writing award, . l ibrary it departments have an obligation to provide reliable services both during and after normal busi- ness hours. the it industry has developed guide- lines for the management of it services, but the library community has been slow to adopt these practices. the delay may be attributed to a number of factors, including a dependence on vendors and consultants for technical expertise, a reliance on librarians who have little formal training in it best practices, and a focus on automation systems instead of infrastructure. larger systems that employ dedicated it professionals to manage the orga- nization’s technology resources likely implement best practices as a matter of course and see no need to discuss them within the library community. in the practice of system and network administration, thomas a. limoncelli, christine j. hogan, and strata r. chalup present a comprehensive look at best practices in managing systems and networks. early in the book they provide a short list of first steps toward improving it ser- vices, one of which is the implementation of some form of monitoring. they point out that without monitoring, systems can be down for extended periods before admin- istrators notice or users report the problem. they dedi- cate an entire chapter to monitoring services. in it, they discuss the two primary types of monitoring—real-time monitoring, which provides information on the current state of services, and historical monitoring, which pro- vides long-term data on uptime, use, and performance. while the software discussed in this article provides both types of monitoring, i focus on real-time monitoring and the value of problem identification and notification. service monitoring does not appear frequently in library literature, and what is written often relates to single-purpose custom monitoring. an article in the september issue of ital describes the development and deployment of a wireless network, including a perl script written to monitor the wireless network and asso- ciated services. the script updates a webpage to display the results and sends an e-mail notifying staff of problems. an enterprise monitoring system could perform these tasks and present the results within the context of the complete infrastructure. it would require using advanced features because of the segregation of networks discussed in their article but would require little or no extra effort than it took to write the single-purpose script. dave pattern at the university of huddersfield shared another perl script that monitors opac functionality. again, the script provided a single-purpose monitoring solution that could be integrated within a larger model. below, i discuss how i modified his script to provide more meaningful monitoring of our opac than the stock webpage monitoring plug-in included with our open- source networks monitoring system, nagios. service monitoring can consist of a variety of tests. in its simplest form, a ping test will verify that a host (server or device) is powered on and successfully con- nected to the network. feher and sondag used ping tests to monitor the availability of the routers and access points on their network, as do i for monitoring connectivity to remote locations. a slightly more meaningful check would test for the establishment of a connection on a port. feher and sondag used this method to check the daemons in their network. the step further would be to evaluate a service response, for example checking the status code returned by a web server. evaluating content forms the next level of meaning. limoncelli, hogan, and chalup discuss end-to-end monitoring, where the moni- toring system actually performs meaningful transactions and evaluates the results. pattern’s script, mentioned above, tests opac func- tionality by submitting a known keyword search and evaluating the response. i implemented this after an incident where nagios failed to alert me to a problem with the opac. the web server returned a status code of to the request for the search page. users, however, want more from an opac, and attempts to search were unsuccessful because of problems with the index server. modifying pattern’s original script, i was able to put together a custom check command that verifies a greater level of functionality by evaluating the number of results for the known search. n software selection limoncelli, hogan, and chalup do not address specific t. michael silver (michael.silver@ualberta.ca) is an mlis stu- dent, school of library and information studies, university of al- berta, edmonton, alberta, canada. monitoring network and service availability with open-source software | silver how-to issues and rarely mention specific products. their book provides the foundational knowledge necessary to identify what must be done. in terms of monitoring, they leave the selection of an appropriate tool to the reader. myriad monitoring tools exist, both commercial and open-source. some focus on network analysis, and some even target specific brands or model lines. the selection of a specific software package should depend on the ser- vices being monitored and the goals for the monitoring. wikipedia lists thirty-five different products, of which eighteen are commercial (some with free versions with reduced functionality or features); fourteen are open- source projects under a general public license or similar license (some with commercial support available but without different feature sets or licenses); and three offer different versions under different licenses. von hagen and jones suggest two of them: nagios and zabbix. i selected the nagios open-source product (http:// www.nagios.org). the software has an established his- tory of active development, a large and active user community, a significant number of included and user- contributed extensions, and multiple books published on its use. commercial support is available from a company founded by the creator and lead developer as well as other authorized solution providers. monitoring appliances based on nagios are available, as are sensors designed to interoperate with nagios. because of the flexibility of a software design that uses a plug-in archi- tecture, service checks for library-specific applications can be implemented. if a check or action can be scripted using practically any protocol or programming language, nagios can monitor it. nagios also provides a variety of information displays, as shown in appendixes a–e. n installation the nagios system provides an extremely flexible solu- tion to monitor hosts and services. the object-orientation and use of plug-ins allows administrators to monitor any aspect of their infrastructure or services using standard plug-ins, user-contributed plug-ins, or custom scripts. additionally, the open-source nature of the package allows independent development of extensions to add features or integrate the software with other tools. community sites such as monitoringexchange (formerly nagios exchange), nagios community, and nagios wiki provide repositories of documentation, plug-ins, extensions, and other tools designed to work with nagios. but that flexibility comes at a cost—nagios has a steep learning curve, and user- contributed plug-ins often require the installation of other software, most notably perl modules. nagios runs on a variety of linux, unix, and berkeley software distribution (bsd) operating systems. for testing, i used a standard linux server distribution installed on a virtual machine. virtualization provides an easy way to test software, especially if an alternate operating system is needed. if given sufficient resources, a virtual machine is capable of running the production instance of nagios. after installing and updating the operating system, i installed the following packages: n apache web server n perl n gd development library, needed to produce graphs and status maps n libpng-devel and libjpeg-devel, both needed by the gd library n gcc and gnu make, which are needed to compile some plug-ins and perl modules most major linux and bsd distributions include nagios in their software repositories for easy instal- lation using the native package management system. although the software in the repositories is often not the most recent version, using these repositories simplifies the installation process. if a reasonably recent version of the software is available from a repository, i will install from there. some software packages are either outdated or not available, and i manually install these. detailed installation instructions are available on the nagios web- site, in several books, and on the previously mentioned websites. the documentation for version includes a number of quick-start guides. most package managers will take care of some of the setup, including modifying the apache configuration file to create an alias available at http://server.name/nagios. i prepared the remainder of this article using the latest stable versions of nagios ( . . ) and the plug-ins ( . . ) at the time of writing. n configuration nagios configuration relies on an object model, which allows a great deal of flexibility but can be complex. planning your configuration beforehand is highly recom- mended. nagios has two main configuration files, cgi.cfg and nagios.cfg. the former is primarily used by the web inter- face to authenticate users and control access, and it defines whether authentication is used and which users can access what functions. the latter is the main configuration file and controls all other program operations. the cfg_file and cfg_dir directives allow the configuration to be split into manageable groupsusing additional recourse files and the object definition files (see figure ). the flexibility offered allows a variety of different structures. i group network information technology and libraries | march devices into groups but create individual files for each server. nagios uses an object- oriented design. the objects in nagios are dis- played in table . a complete review of nagios configuration is beyond the scope of this article. the documenta- tion installed with nagios covers it in great detail. special attention should be paid to the concepts of templates and object inheritance as they are vital to creating a man- ageable configuration. the discussion below provides a brief introduction, while appendixes f–j provide concrete examples of working configuration files. n cgi.cfg the cgi.cfg file controls the web interface and its asso- ciated cgi (common gateway interface) programs. during testing, i often turn off authentication by setting use_authentication to if the web interface is not accessible from the internet. there also are various configuration directives that provide greater control over which users can access which features. the users are defined in the /etc/nagios/htpasswd.users file. a summary of com- mands to control entries is presented in table . the web interface includes other features, such as sounds, status map displays, and integration with other products. discussion of these directives is beyond the scope of this article. the cgi.cfg file provided with the software is well commented, and the nagios documen- tation provides additional information. a number of screenshots from the web interface are provided in the appendixes, including status displays and reporting. n nagios.cfg the nagios.cfg file controls the operation of everything except the web interface. although it is possible to have a single monolithic configuration file, organizing the con- figuration into manageable files works better. the two main directives of note are cfg_file, which defines a single file that should be included, and cfg_dir, which includes all files in the specified directory with a .cfg extension. a third type of file that gets included is resource.cfg, which defines various macros for use in commands. organizing the object files takes some thought. i monitor more than one hundred services on roughly seventy hosts, so the method of organizing the files was of more than academic interest. i use the following con- figuration files: n commands.cfg, containing command definitions n contacts.cfg, containing the list of contacts and associated information, such as e-mail address, (see appendix h) n groups.cfg, containing all groups—hostgroups, ser- vicegroups, and contactgroups, (see appendix g) n templates.cfg, containing all object templates, (see appendix f) n timeperiods.cfg, containing the time ranges for checks and notifications all devices and servers that i monitor are placed in directories using the cfg_dir directive: servers—contains server configurations. each file includes the host and service configurations for a physical or virtual server. devices—contains device information. i create indi- vidual files for devices with service monitoring that goes beyond simple ping tests for connectiv- table . nagios objects object used for hosts servers or devices being monitored hostgroups groups of hosts services services being monitored servicegroups groups of services timeperiods scheduling of checks and notifications commands checking hosts and services notifying contacts processing performance data event handling contacts individuals to alert contactgroups groups of contacts figure . nagios configura- tion relationships. copyright © ethan galstead, nagios enterprises. used with permis- sion. monitoring network and service availability with open-source software | silver ity. devices monitored solely for connectivity are grouped logically into a single file. for example, we monitor connectivity with fifty remote locations, and all fifty of them are placed in a single file. the resource.cfg file uses two macros to define the path to plug-ins and event handlers. thirty other macros are available. because the cgi programs do not read the resource file, restrictive permissions can be applied to them, enabling some of the macros to be used for user- names and passwords needed in check commands. placing sensitive information in service configurations exposes them to the web server, creating a security issue. n configuration the appendixes include the object configuration files for a simple monitoring situation. a switch is monitored using a simple ping test (see appendix j), while an opac server on the other side of the switch is monitored for both web and z . operations (see appendix i). note that the opac configuration includes a parents directive that tells nagios that a problem with the gateway-switch will affect connectivity with the opac server. i monitor fifty remote sites. if my router is down, a single notification regarding my router provides more information if it is not buried in a storm of notifications about the remote sites. the web port, web service, and opac search services demon- strate different levels of monitoring. the web port simply attempts to establish a connection to port without evalu- ating anything beyond a successful connection. the web service check requests a specific page from the web server and evaluates only the status code returned by the server. it displays a warning because i configured the check to download a file that does not exist. the web server is run- ning because it returns an error code, hence the warning status. the opac search uses a known search to evaluate the result content, specifically whether the correct number of results is returned for a known search. i used a number of templates in the creation of this configuration. templates reduce the amount of repeti- tive typing by allowing the reuse of directives. templates can be chained, as seen in the host templates. the opac definition uses the linux-server template, which in turn uses the generic-host template. the host definition inher- its the directives of the template it uses, overriding any elements in both and adding new elements. in practical terms, generic-host directives are read first. linux-server directives are applied next. if there is a conflict, the linux- server directive takes precedence. finally, opac is read. again, any conflicts are resolved in favor of the last con- figuration read, in this case opac. n plug-ins and service checks the nagios plugins package provides numerous plug-ins, including the check-host-alive, check_ping, check_tcp, and check_http commands. using the plug-ins is straightfor- ward, as demonstrated in the appendixes. most plug- ins will provide some information on use if executed with—help supplied as an argument to the command. by default, the plug-ins are installed in /usr/lib/nagios/ plugins. some distributions may install them in a differ- ent directory. the plugins folder contains a subfolder with user- contributed scripts that have proven useful. most of these plug-ins are perl scripts, many of which require additional perl modules available from the comprehensive perl archive network (cpan). the check_hip_search plug-in (appendix k) used in the exam- ples requires additional modules. installing perl mod- ules is best accomplished using the cpan perl module. detailed instructions on module installation are avail- able online. some general tips: n gcc and make should be installed before trying to install perl modules, regardless of whether you are installing manually or using cpan. most modules are provided as source code, which may require compiling before use. cpan automates this pro- cess but requires the presence of these packages. n alternately, many linux distributions provide perl module packages. using repositories to install usu- ally works well assuming the repository has all the needed modules. in my experience, that is rarely the case. table . sample commands for managing the htpasswd.users file create or modify an entry, with password entered at a prompt: htpasswd /etc/nagios/htpasswd.users <username> create or modify an entry using password from the command line: htpasswd -b /etc/nagios/htpasswd.users <username> <password> delete an entry from the file: htpasswd -d /etc/nagios/htpasswd.users <username> information technology and libraries | march n many modules depend on other modules, some- times requiring multiple install steps. both cpan and distribution package managers usually satisfy dependencies automatically. manual installation requires the installer to satisfy the dependencies one by one. n most plug-ins provide information on required software, including modules, in a readme file or in the source code for the script. in the absence of such documentation, running the script on the command line usually produces an error contain- ing the name of the missing module. n testing should be done using the nagios user. using another user account, especially the root user, to create directories, copy files, and run programs creates folders and files that are not accessible to the nagios user. the best practice is to use the nagios user for as much of the configuration and testing as possible. the lists and forums frequently include questions from new users that have successfully installed, configured, and tested nagios as the root user and are confused when nagios fails to start or function properly. n advanced topics once the system is running, more advanced features can be explored. the documentation describes many such enhancements, but the following may be particularly use- ful depending on the situation. n nagios provides access control through the combi- nation of settings in the cgi.cfg and htpasswd.users files. library administration and staff, as well as patrons, may appreciate the ability to see the sta- tus of the various systems. however, care should be taken to avoid disclosing sensitive information regarding the network or passwords, or allowing access to cgi programs that perform actions. n nagios permits the establishment of dependency relationships. host dependencies may be useful in some rare circumstances not covered by the parent–child relationships mentioned above, but service dependencies provide a method of connect- ing services in a meaningful manner. for example, certain opac functions are dependent on ils ser- vices. defining these relationships takes both time and thought, which may be worthwhile depending on any given situation. n event handlers allow nagios to initiate certain actions after a state change. if nagios notices that a particular service is down, it can run a script or program to attempt to correct the problem. care should be taken when creating these scripts as ser- vice restarts may delete or overwrite information critical to solving a problem, or worsen the actual situation if an attempt to restart a service or reboot a server fails. n nagios provides notification escalations, permit- ting the automatic notification of problems that last longer than a certain time. for example, a service escalation could send the first three alerts to the admin group. if properly configured, the fourth alert would be sent to the managers group as well as the admin group. in addition to escalating issues to management, this feature can be used to establish a series of responders for multiple on-call personnel. n nagios can work in tandem with remote machines. in addition to custom scripts using secure shell (ssh), the nagios remote plug-in executor (nrpe) add-on allows the execution of plug-ins on remote machines, while the nagios service check acceptor (nsca) add-on allows a remote host to submit check results to the nagios server for processing. implementing nagios on the feher and sondag wireless network mentioned earlier would require one of these options because the wireless network is not accessible from the external network. these add-ons also allow for distributed monitoring, sharing the load among a number of servers while still providing the administrators with a single interface to the entire monitored network. the nagios exchange (http://exchange.nagios .org/) contains similar user-contributed programs for windows. n nagios can be configured to provide redundant or failover monitoring. limoncelli, hogan, and chalup call this metamonitoring and describe when it is needed and how it can be implemented, suggesting self-monitoring by the host or having a second monitoring system that only monitors the main system. nagios permits more complex configurations, allowing for either two servers operating in parallel, only one of which sends notifications unless the main server fails, or two servers communicating to share the monitoring load. n alternative means of notification increase access to information on the status of the network. i imple- mented another open-source software package, quickpage, which allows nagios text messages to be sent from a computer to a pager or cell phone. appendix l shows a screenshot of a firefox exten- sion that displays host and service problems in the status bar of my browser and provides optional audio alerts. the nagios community has devel- oped a number of alternatives, including special- ized web interfaces and rss feed generators. monitoring network and service availability with open-source software | silver n appropriate use monitoring uses bandwidth and adds to the load of machines being monitored. accordingly, an it depart- ment should only monitor its own servers and devices, or those for which it has permission to do so. imagine what would happen if all the users of a service such as worldcat started monitoring it! the additional load would be noticeable and could conceivably disrupt service. aside from reasons connected with being a good “netizen,” monitoring appears similar to port-scanning, a technique used to discover network vulnerabilities. an organization that blithely monitors devices without the owner’s permission may find their traffic is throttled back or blocked entirely. if a library has a definite need to moni- tor another service, obtaining permission to do so is a vital first step. if permission is withheld, the service level agree- ment between the library and its service provider or ven- dor should be reevaluated to ensure that the provider has an appropriate system in place to respond to problems. n benefits the system-administration books provide an accurate overview of the benefits of monitoring, but personally reaping those benefits provides a qualitative background to the experience. i was able to justify the time spent on setting up monitoring the first day of production. one of the available plug-ins monitors sybase database servers. it was one of the first contributed plug-ins i implemented because of past experiences with our production database running out of free space, causing the system to become nonfunctional. this happened twice, approximately a year apart. each time, the integrated library system was down while the vendor addressed the issue. when i enabled the sybase service checks, nagios immediately returned a warning for the free space. the advance warning allowed me to work with the vendor to extend the database volume with no downtime for our users. that single event con- vinced the library director of the value of the system. since that time, nagios has proven its worth in alert- ing it staff to problem situations, providing information on outage patterns both for in-house troubleshooting and discussions with service providers. n conclusion monitoring systems and services provides it staff with a vital tool in providing quality customer service and managing systems. installing and configuring such a system involves a learning curve and takes both time and computing resources. my experiences with nagios have convinced me that the return on investment more than justifies the costs. references . thomas a. limoncelli, christina j. hogan, and strata r. chalup, the practice of system and network administration, nd ed. (upper saddle river, n.j.: addison-wesley, ): . . ibid., – . . james feher and tyler sondag, “administering an open- source wireless network,” information technology & libraries , no. (sept. ): – . . dave pattern, “keeping an eye on your hip,” online post- ing, jan. , , self-plagiarism is style, http://www.daveyp .com/blog/archives/ (accessed nov. , ). . feher and sondag, “administering an open-source wire- less network,” – . . ibid., , – . . limoncelli, hogan, and chalup, the practice of system and network administration, – . . pattern, “keeping an eye on your hip.” . limoncelli, hogan, and chalup, the practice of system and network administration, xxv. . “comparison of network monitoring systems,” wikipe- dia, the free encyclopedia, dec. , , http://en.wikipedia .org/wiki/comparison_of_network_monitoring_systems (accessed dec. , ). . william von hagen and brian k. jones, linux server hacks, vol. (sebastopol, calif.: o’reilly, ): – (zabbix), – (nagios). . monitoringexchange, http://www.monitoringexchange. org/ (accessed dec. , ); nagios community, http:// community.nagios.org (accessed dec. , ); nagios wiki, http://www.nagioswiki.org/ (accessed dec. , ). . “nagios documentation,” nagios, mar. , , http:// www.nagios.org/docs/ (accessed dec. , ); david joseph- sen, building a monitoring infrastructure with nagios (upper saddle river, n.j.: prentice hall, ); wolfgang barth, nagios: system and network monitoring, u.s. ed. (san francisco: open source press; no starch press, ). . ethan galstead, “nagios quickstart installation guides,” nagios .x documentation, nov. , , http://nagios.source forge.net/docs/ _ /quickstart.html (accessed dec. , ). . the perl directory, (http://www.perl.org/) contains com- plete information on perl. specific information on using cpan is available in “how do i install a module from cpan?” perlfaq , nov. , , http://perldoc.perl.org/perlfaq .html (accessed dec. , ). . limoncelli, hogan, and chalup, the practice of system and network administration, – . . thomas dwyer iii, qpage solutions, http://www.qpage .org/ (accessed dec. , ). . petr Šimek, “nagioschecker,” google code, aug. , , http://code.google.com/p/nagioschecker/ (accessed dec. , ). . “notifications,” monitoringexchange, http://www .monitoringexchange.org/inventory/utilities/addon-proj- ects/notifications (accessed dec. , ). information technology and libraries | march appendix a. service detail display from test system appendix b. service details for opac (hip) and ils (horizon) servers from production system appendix c. sybase freespace trends for a specified period appendix d. connectivity history for a specified period appendix e. availability report for host shown in appendix d appendix f. templates.cfg file ############################################################################ # templates.cfg - sample object templates ############################################################################ ############################################################################ # contact templates ############################################################################ monitoring network and service availability with open-source software | silver # generic contact definition template - this is not a real contact, just # a template! define contact{ name generic-contact service_notification_period x host_notification_period x service_notification_options w,u,c,r,f,s host_notification_options d,u,r,f,s service_notification_commands notify-service-by-email host_notification_commands notify-host-by-email register } ############################################################################ # host templates ############################################################################ # generic host definition template - this is not a real host, just # a template! define host{ name generic-host notifications_enabled event_handler_enabled flap_detection_enabled failure_prediction_enabled process_perf_data retain_status_information retain_nonstatus_information notification_period x register } # linux host definition template - this is not a real host, just a template! define host{ name linux-server use generic-host check_period x check_interval retry_interval max_check_attempts check_command check-host-alive notification_period workhours notification_interval notification_options d,u,r contact_groups admins register } appendix f. templates.cfg file (cont.) information technology and libraries | march # define a template for switches that we can reuse define host{ name generic-switch use generic-host check_period x check_interval retry_interval max_check_attempts check_command check-host-alive notification_period x notification_interval notification_options d,r contact_groups admins register } ############################################################################ # service templates ############################################################################ # generic service definition template - this is not a real service, # just a template! define service{ name generic-service active_checks_enabled passive_checks_enabled parallelize_check obsess_over_service check_freshness notifications_enabled event_handler_enabled flap_detection_enabled failure_prediction_enabled process_perf_data retain_status_information retain_nonstatus_information is_volatile check_period x max_check_attempts normal_check_interval retry_check_interval contact_groups admins notification_options w,u,c,r notification_interval notification_period x register } appendix f. templates.cfg file (cont.) monitoring network and service availability with open-source software | silver # define a ping service. this is not a real service, just a template! define service{ use generic-service name ping-service notification_options n check_command check_ping! . , %! . , % register } appendix f. templates.cfg file (cont.) appendix g. groups.cfg file ############################################################################ # contact group definitions ############################################################################ # we only have one contact in this simple configuration file, so there is # no need to create more than one contact group. define contactgroup{ contactgroup_name admins alias nagios administrators members nagiosadmin } ############################################################################ # host group definitions ############################################################################ # define an optional hostgroup for linux machines define hostgroup{ hostgroup_name linux-servers ; the name of the hostgroup alias linux servers ; long name of the group } # create a new hostgroup for ils servers define hostgroup{ hostgroup_name ils-servers ; the name of the hostgroup alias ils servers ; long name of the group } # create a new hostgroup for switches define hostgroup{ hostgroup_name switches ; the name of the hostgroup alias network switches ; long name of the group } ############################################################################ # service group definitions ############################################################################ information technology and libraries | march # define a service group for network connectivity define servicegroup{ servicegroup_name network alias network infrastructure services } # define a servicegroup for ils define servicegroup{ servicegroup_name ils-services alias ils related services } appendix g. groups.cfg file (cont.) appendix h. contacts.cfg ############################################################################ # contacts.cfg - sample contact/contactgroup definitions ############################################################################ # just one contact defined by default - the nagios admin (that’s you) # this contact definition inherits a lot of default values from the # ‘generic-contact’ template which is defined elsewhere. define contact{ contact_name nagiosadmin use generic-contact alias nagios admin email nagios@localhost } appendix i. opac.cfg ############################################################################ # opac server ############################################################################ ############################################################################ # host definition ############################################################################ # define a host for the server we’ll be monitoring # change the host_name, alias, and address to fit your situation define host{ use linux-server host_name opac parents gateway-switch alias opac server monitoring network and service availability with open-source software | silver appendix i. opac.cfg (cont.) address . . . } ############################################################################ # service definitions ############################################################################ # create a service for monitoring the http port define service{ use generic-service host_name opac service_description web port check_command check_tcp! } # create a service for monitoring the web service define service{ use generic-service host_name opac service_description web service check_command check_http!-u/bogusfilethatdoesnotexist.html } # create a service for monitoring the opac search define service{ use generic-service host_name opac service_description opac search check_command check_hip_search } # create a service for monitoring the z . port define service{ use generic-service host_name opac service_description z port check_command check_tcp! } appendix j. switches.cfg ############################################################################ # switch.cfg - sample config file for monitoring switches ############################################################################ ############################################################################ # host definitions ############################################################################ information technology and libraries | march appendix k. check_hip_search script #!/usr/bin/perl -w ######################### # check horizon information portal (hip) status. # hip is the web-based interface for dynix and horizon # ils systems by sirsidynix corporation. # # this plugin is based on a standalone perl script written # by dave pattern. please see # http://www.daveyp.com/blog/index.php/archives/ / # for the original script. # # the original script and this derived work are covered by # http://creativecommons.org/licenses/by-nc-sa/ . / ######################### use strict; use lwp::useragent; # note the requirement for perl module lwp::useragent! use lib “/usr/lib/nagios/plugins”; use utils qw($timeout %errors); # define the switch that we’ll be monitoring define host{ use generic-switch host_name gateway-switch alias gateway switch address . . . hostgroups switches } ############################################################################ ### # service definitions ############################################################################ ### # create a service to ping to switches # note this entry will ping every host in the switches hostgroup define service{ use ping-service hostgroups switches service_description ping normal_check_interval retry_check_interval } appendix j. switches.cfg monitoring network and service availability with open-source software | silver ### some configuration options my $hipserverhome = “http://ipac.prl.ab.ca/ipac /ipac. jsp?profile=alap”; my $hipserversearch = “http://ipac.prl.ab.ca/ipac /ipac.jsp?menu=se arch&aspect=subtab &npp= &ipp= &spp= &profile=alap&ri=&index=.gw&term=li nux&x= &y= &aspect=subtab &getxml=true”; my $hipsearchtype = “xml”; my $httpproxy = ‘’; ### check home page is available... { my $ua = lwp::useragent->new; $ua->timeout( ); if( $httpproxy ) { $ua->proxy( ‘http’, $httpproxy ) } my $response = $ua->get( $hipserverhome ); my $status = $response->status_line; if( $response->is_success ) { } else { print “hip_search critical: $status\n”; exit $errors{‘critical’}; } } ### check search page is returning results... { my $ua = lwp::useragent->new; $ua->timeout( ); if( $httpproxy ) { $ua->proxy( ‘http’, $httpproxy ) } my $response = $ua->get( $hipserversearch ); my $status = $response->status_line; if( $response->is_success ) { my $results = ; my $content = $response->content; if( lc( $hipsearchtype ) eq ‘html’ ) { if ( $content =~ /\<b\>(\d+?)\<\/b\>\ \;titles matched/ ) { $results = $ ; appendix k. check_hip_search script (cont.) information technology and libraries | march } } if( lc( $hipsearchtype ) eq ‘xml’ ) { if( $content =~ /\<hits\>(\d+?)\<\/hits\>/ ) { $results = $ ; } } ### modified section - original script triggered another function to ### save results to a temp file and email an administrator. unless( $results ) { print “hip_search critical: no results returned|results= \n”; exit $errors{‘critical’}; } if ( $results ) { print “hip_search ok: $results results returned|results=$results\n”; exit $errors{‘ok’}; } } } appendix k. check_hip_search script (cont.) appendix l. nagios checker display editorial board thoughts | dehmlow mark dehmloweditorial board thoughts the ten commandments of interacting with nontechnical people m ore than ten years of working with technology and interacting with nontechnical users in a higher education environment has taught me many lessons about successful communication strategies. somehow, in that time, i have been fortunate to learn some effective mechanisms for providing constructive support and leading successful technical projects with both technically and “semitechnically” minded patrons and librarians. i have come to think of myself as some- one who lives in the “in between,” existing more in the beyond than the bed or the bath, and, while not a native of either place, i like to think that i am someone who is comfortable in both the technical and traditional cliques within the library. ironically, it turns out that the most critical pieces to successfully implementing technology solutions and bridging the digital divide in libraries has been categorically nontechnical in nature; it all comes down to collegiality, clear communication, and a commit- ment to collaboration. as i ruminated on the last ten plus years of work- ing in technology, i began to think of the behaviors and techniques that have proved most useful in developing successful relationships across all areas of the library. the result is this list of the top ten dos and don’ts for those of us self-identified techies who are working more and more often with the self-identified nontechnical set. . be inclusive—i have been around long enough to see how projects that include only technical people are doomed to scrutiny and criticism. the single best strategy i have found to getting buy-in for technical projects is to include key stakeholders and those with influence in project planning and core decision-making. not only does this create support for projects, but it encourages others to have a sense of ownership in project implementa- tion—and when people feel ownership for a proj- ect, they are more likely to help it succeed. . share the knowledge—i don’t know if it is just the nature of librarianship, but librarians like to know things, and more often than not they have a healthy sense of curiosity about how things work. i find it goes a long way when i take a few moments to explain how a particular technology works. our public services specialists, in particular, often want to know the details of how our digital tools work so that they can teach users most effectively and answer questions users have about how they func- tion. sharing expertise is a really nice way to be inclusive. . know when you have shared enough—in the same way that i don’t need to know every deep detail of collections management to appreciate it, most nontechies don’t need hour-long lectures on how each component of technology relates to the other. knowing how much information to share when describing concepts is critical to keeping people’s interest and generally keeping you approachable. . communicate in english—it is true that every spe- cialization has its own vocabulary and acronyms (oh how we love acronyms in libraries) that have no relevance to nonspecialists. i especially see this in the jargon we use in the library to describe our tools and services. the best policy is to avoid jar- gon and explain concepts in lay-person’s terms or, if using jargon is unavoidable, define specialized words in the simplest terms possible. using analo- gies and drawing pictures can be excellent ways to describe technical concepts and how they work. it is amazing how much from kindergarten remains relevant later in life! . avoid techno-snobbery—i know that i am risking vir- tual ostracism in writing this, but i think it needs to be said. just because i understand technology does not make me better than others, and i have heard some variant of the “cup holder on the computer” joke way too often. even if you don’t make these kinds of comments in front of people who aren’t as technically capable as you, the attitude will be apparent in your interactions, and there is truly nothing more condescending. . meet people halfway—when people are trying to ask technology-related questions or converse about technical issues, don’t correct small mistakes. instead, try to understand and coax out their mean- ing; elaborate on what they are saying, and extend the conversation to include information they might not be aware of. people don’t like to be corrected or made to feel stupid—it is embarrassing. if their understanding is close enough to the basic idea, letting small mistakes in terminology slide can create an opening for a deeper understanding. you can provide the correct terminology when talking about the topic without making a point to correct people. . don’t make a clean technical/nontechnical distinction— after once offering the “technical” perspective on a topic, one librarian said to me that it wasn’t that they themselves didn’t have any technical mark dehmlow (mdehmlow@nd.edu) is digital initiatives librarian, hesburgh libraries, university of notre dame, notre dame, indiana. information technology and libraries | june perspective, it just wasn’t perhaps as extensive as mine. each person has some level of technical expertise; it is better to encourage the development of that understanding rather than compartmental- izing people on the basis of their area of expertise. . don’t expect everyone to be interested—just because i chose a technical track and am interested in it doesn’t mean everyone should be. sometimes peo- ple just want to focus on their area of expertise and let the technical work be handled by the techies. . assume everyone is capable—at least at some level. sometimes it is just a question of describing con- cepts in the right way, and besides, not every- one should be a programmer. everyone brings their own skills to the table and that should be respected. . expertise is just that—and no one, no one knows everything. there just isn’t enough time, and our brains aren’t that big. embrace those with different expertise, and bring those perspectives into your project planning. a purely technical perspective, while perhaps being efficient, may not provide a practical or intuitive solution for users. diversity in perspective creates stronger projects. in the same way that the most interesting work in academia is becoming increasingly more multidisci- plinary, so too the most successful work in libraries needs to bring diverse perspectives to the fore. while it is easy to say libraries are constantly becoming more technically oriented because of the expanse of digital collections and services, the need for the convergence of the technical and traditional domains is clear—digital preservation is a good example of an area that requires the lessons and strengths learned from physical preservation, and, if any- thing, the technical aspects still raise more questions than solutions—just see henry newman’s article “rocks don’t need to be backed up” to see what i mean. increasingly, as we develop and implement applications that better leverage our collections and highlight our services, their success hinges on their usability, user-driven design, and implementations based on user feedback. these “user”-based evaluation techniques fit more closely with traditional aspects of public services: interacting with patrons. lastly, it is also important to remember that technol- ogy can be intimidating. it has already caused a good deal of anxiety for those in libraries who are worried about long-term job security as technology continues to initiate changes in the way we perform our jobs. one of the best ways to bring people along is to demystify the scary parts of technology and help them see a role for themselves in the future of the library. going back to maslow’s hierar- chy of needs, people want to feel a sense of security and belonging, and i believe it is incumbent upon those of us with a deep understanding of technology to help bring the technical to the traditional in a way that serves every- one in the process. reference . henry newman, “rocks don’t need to be backed up,” enterprise storage forum.com (mar. , ), www.enterprise storageforum.com/continuity/features/article.php/ (accessed april , ). public access technologies in public libraries | bertot john carlo bertot public access technologies in public libraries: effects and implications public libraries were early adopters of internet-based technologies and have provided public access to the internet and computers since the early s. the landscape of public-access internet and computing was substantially different in the s as the world wide web was only in its initial development. at that time, public libraries essentially experimented with public- access internet and computer services, largely absorbing this service into existing service and resource provision without substantial consideration of the management, facilities, staffing, and other implications of public-access technology (pat) services and resources. this article explores the implications for public libraries of the provi- sion of pat and seeks to look further to review issues and practices associated with pat provision resources. while much research focuses on the amount of public access that public libraries provide, little offers a view of the effect of public access on libraries. this article provides insights into some of the costs, issues, and challenges associated with public access and concludes with recommendations that require continued exploration. p ublic libraries were early adopters of internet-based technologies and have provided public access to the internet and computers since the early s. in , . percent of public libraries were connected to the internet, and . percent offered public-access com- puters. by , internet connectivity in public libraries grew to . percent, and . percent of public librar- ies provided public internet access. the landscape of public-access internet and computing was substantially different in the s, as the world wide web was only in its initial development. at that time, public libraries essentially experimented with public-access internet and computer services, largely absorbing this service into existing service and resource provision without substan- tial consideration of the management, facilities, staffing, and other implications of public-access technology (pat) services and resources. using case studies conducted at thirty-five public libraries in five geographically dispersed and demograph- ically diverse states, this article explores the implications for public libraries of the provision of pat. the researcher also conducted interviews with state library agency staff prior to visiting libraries in each state. the goals of this article are to n explore the level of support pat requires within public libraries; n explore the implications of pat on public libraries, including management, building planning, staff- ing, and other support issues; n explore current pat support practices; n identify issues and challenges public libraries face in maintaining and supporting their pat infra- structure; and n identify factors that contribute to successful pat practices. this article seeks to look beyond the provision of pat by public libraries and review issues and practices associated with pat–provision resources. while much research focuses on the amount of public access that public libraries provide, little offers a view of the effect of public access on libraries. this article provides insights into some of the costs, issues, and challenges associated with public access, and it concludes with recommenda- tions that require continued exploration. n literature review quickly over time, public libraries increased their pub- lic-access provision substantially (see figures and ). connectivity grew from . percent in to nearly percent in . moreover, nearly all libraries that connected to the internet offered public-access internet services. simultaneously, the average number of public- access computers grew from . per public library in to per public library in . accompanying and in support of the continual growth of basic connec- tivity and computing infrastructure was a demand for broadband connectivity. indeed, since , connectiv- ity has progressed from dial-up phone lines to leased lines and other forms of high-speed connectivity. the extent of the growth in public-access services within public libraries is profound and substantive, leading to the development of new internet-based service roles for public libraries. and public access to the internet through public libraries provides a number of com- munity benefits to different populations within served communities. overlaid onto the public-access infrastructure is an increasingly complex service mix that now includes access to digital content (e.g., databases and digital john carlo bertot (jbertot@umd.edu) is professor and director of the center for library innovation in the college of information studies at the university of maryland, college park. information technology and libraries | june libraries), integrated library systems (ilss), voice over internet protocol (voip), digital reference, and a host of other services and resources—some for public access, others for back-office library operations. and patrons do use these services in increasing amounts—both in the library and in everyday life. in fact, . percent of public libraries report that they do not have an adequate number of public-access computers some or all of the time and have resorted to time limits and wireless access to extend public-access services. by , as connectivity and public-access computer infrastructure grew, so ensued the need to provide a range of publicly available services and resources: n . percent of public libraries provide access to licensed databases n . percent of public libraries offer technology training n . percent of public libraries provide e-govern- ment services (e.g., locating government infor- mation and helping patrons complete online applications) n . percent of public libraries provide digital refer- ence services n . percent of public libraries offer access to e-books the list is not exhaustive, but illustrative, since librar- ies do offer other services such access to homework resources, video content, audio content, and digitized collections. as public libraries expanded these services, man- agement realized that they needed to plan and evalu- ate technology-based services. over the years, a range of technology management, planning, and evaluation resources emerged to help public libraries cope with their technology-based resources—those both publicly avail- able and for administrative operations. but increasingly, public libraries report the strain that pat services promulgate. this centers on four key areas: n maintenance and management. the necessary main- tenance and management requirements of pat places an additional burden on existing staff, many of whom do not possess technology expertise to troubleshoot, fix, and support internet-based ser- vices and resources that patrons access. n staff. libraries consistently cite staff expertise and availability as a barrier to the addition, support, and management of pat. indeed, as described in previous sections, some libraries have experienced a decline in library staff. n finances. there is evidence of stagnant funding for libraries at the local level as well as a shift in expen- ditures from staff and collections to operational costs such as utilities and maintenance. n buildings. the buildings are inadequate in terms of space and infrastructure (e.g., wiring and cabling) to support additional public access. this article explores these four areas through a site- visit method in an effort to go beyond a quantitative assessment of pat within the public library community. though related in terms of topic area and author, this study was conducted separately from the public library internet surveys conducted since and offers insights into the provision of pat services and resources that a national survey cannot explore in such depth. figure . public-access internet connectivity from through figure . public-access internet workstations from through public access technologies in public libraries | bertot n method the researcher visited thirty-five public libraries in five geographically and demographically diverse states between october and may . the states were in the west, southwest, southeast, and mid-atlantic regions. the libraries visited included urban, suburban, rural, and native american public libraries that served populations ranging from a few hundred to more than half a million. the communities that the libraries served varied in terms of poverty, race, income, age, employment, and education demographics. prior to visiting the public library sites, the researcher conducted interviews with state library agency staff to better understand the public library con- text within each state and to explore overall pat issues, strategies, and other factors within the state. the following research questions guided the site visits: n what are the community and library contexts in which the library provides pat? n what are the pat services and resources that the library makes available to its community? n what pat services and resources does the library desire to provide to its community? n what is the relationship between provided and desired pat and the effect on the library (e.g., staff, finances, the building, and management)? n what are the perceived benefits to the library and its community gains through pat in the library? n what are the issues and barriers that the library encounters in providing pat services and resources? n how does the library manage and maintain its pat? the researcher visited each library for four to six hours. during that time, he interviewed the library direc- tor and/or branch manager and technology support staff (either a specific library position, designated library employee, or city or county it staff person), toured the library facility, and conducted a brief technology inven- tory. at some libraries, the researcher was able to meet with community partners that in some way collaborated with the library to provide pat services and resources (e.g., educational institutions that collaborated with libraries to provide access to broadband or volunteers who conducted technology training sessions). interviews were recorded and transcribed, and the technology inventories were entered into a microsoft excel spreadsheet for analysis. the transcripts were coded using thematic content analytic schemes to allow for the identification of key issues regarding pat areas. this approach enabled the researcher to use an iterative site-visit strategy that used findings from previous site visits to inform subsequent visits. to ensure valid and reliable data, the researcher used a three-stage strategy: . site-visit reports were completed and sent to th libraries for review. corrections from libraries were incorporated into a final site-visit report. . a final state-based site-visit report was compiled for distribution to state library agency staff and also incorporated their corrections. this provided a state-level reliability and validity check. . a summary of key findings was distributed to six experts in the public library technology environ- ment, three of which were public library technol- ogy managers and three of which were technology consultants who worked with public libraries. in combination, this approach provided three levels of data quality checks, thus providing both internal (library and state) and external (technology expert) support for the findings. the findings in this article are limited to the libraries visited and interviews conducted with public librarians and state library agency staff. however, themes emerged early during the site-visit process and were reinforced through subsequent interviews and visits across the states and libraries visited. in addition, the use of external reviewers of the findings lends additional, but limited, support to the findings. n findings this section presents the results of the site visits and interviews with state library agency staff and public librarians. the article presents the findings by key areas surrounding pat in public libraries. the public-access context public libraries have a range of pat installed in their libraries for patron use. these technologies include pub- lic-access computers, wireless (wifi) access, ilss, online databases, digital reference, downloadable audio and video, and others. many of these services and resources are also available to patrons from outside library build- ings, thus extending the reach (and support issues) of the library beyond the library’s walls. in addition, when libraries do not provide direct access to resources and ser- vices, they serve as access points to those services, such as online gaming and social networking. while libraries can and do deploy a number of technologies for public use, it is possible to group these information technology and libraries | june technologies broadly into two overlapping categories: n hardware. library pat hardware can include pub- lic-access computers, public-access computing reg- istration (i.e., reservation) systems, self-checkout stations, printers, faxes, laptops, and a range of other devices and systems. some of these technolo- gies may have additional devices, such as those required for persons with disabilities. within the hardware grouping are networking technologies that include a range of hardware and software to enable a range of library networks to run (e.g., routers, hubs, switches, telecommunications lines, and networking software). n software. software can include device operating system software (e.g., microsoft windows, mac os, and linux), device application software (e.g., microsoft office, openoffice, graphics software, audio software, e-book readers, assistive software, and others), and functional software (e.g., web browsers, online databases, and digital reference). in short, public libraries make use of a range of tech- nologies that the public uses in some way. each type of technology requires skills, management, implementation, and maintenance, all of which are discussed later. in the building, all of these products and services come together at the library’s public-access computers, or patron mobile device if wifi is available. moreover, patrons increasingly want to use their portable devices (e.g., usb drives, ipods, and others) with library tech- nology. this places pressure on libraries to not just offer public-access computers, but also to support a range of technologies and services. thus the environment in which libraries offer pat is complex and requires substantial technical expertise, support, and maintenance in key areas of applications, computers, and networking. moreover, as discussed below, patrons are increasingly demanding market-based approaches to pat. these demands—which are largely about single-point access to a range of information ser- vices and resources—are often at odds with library tech- nology that is based on stove-piped approaches (e.g, ils, e-books, and licensed resources) and that do not necessar- ily lend themselves to seamless integration. n external pressures on pats the advent and increased use by the public of google, amazon, itunes, youtube, myspace, second life, and other networked services affects public libraries in a number of ways. this article discusses these services and resources from the perspective of an information marketplace of which the public library is one entrant. interviewed librarians overwhelmingly indicated that users now expect library services to resemble those in the marketplace. users expect the look and feel, integration, service capabilities, interactivity, and personalization and customization that they experience while engaging in social networking, online searching, online purchas- ing, or other online activities. and within the library building, patrons expect the services to integrate at the public-access computer entry point—not distributed throughout the library in a range of locations, worksta- tions, or devices. said differently, they expect to have a “mylibrary.com” experience that allows for seamless integration across the library’s services but also facilitates the use of personal technologies (e.g., ipods, mp players, and usb devices). thus users expect the library’s services to resemble those services offered by a range of informa- tion service providers. importantly, however, librarians indicated that library systems on which their services and resources reside by and large do not integrate seamlessly—nor were they designed to do so. public-access computers are gateways to the internet; the ils exists for patrons to search for and locate library holdings; and online databases, e-books, audiobooks, etc., are extensions of the library’s holdings but are not physical items under a library’s control and thus subject to a vendor’s information and business mod- els. while library vendors and the library community are working to develop more integrated products that lead users to the information they seek, the technology is under development. there are three significant issues that libraries face because of market pressures: ( ) the pressures all come together at a single point—the public-access computer; ( ) users want a customized experience while using tech- nology designed for the general public, not the individual user; and ( ) users have choices in the information mar- ketplace. one participant indicated, “if the library cannot match what users have access to on the outside, users will and do move on.” managing and maintaining public access managing the public-access computer environment for public libraries is an growing challenge. there are a num- ber of management areas with which public librarians contend: n public-access computers—the computers and laptops (if applicable) themselves, which can include any- thing from keyboards and mice to troubleshooting a host of computer problems (it is important to note that these may be computers that often vary in age and composition, come from a range of ven- dors, run different operating systems, and often public access technologies in public libraries | bertot have different application software versions). n peripheral management—the printers, faxes, scan- ners, and other equipment that are part of the library’s overall public access infrastructure. n public-access management software or systems—these may include online or in-building computer-based reservations (which encompasses specialized reser- vations such as teen machines, gaming computers, computers for seniors, and so on), time manage- ment (set to the library’s decided-upon time allot- ment), filtering, security, logins, virtual machines, etc. n wireless access—this may include logins and config- urations for patrons to gain access to the library’s wireless network. n bandwidth management—this may include the need to allocate bandwidth differently as needs increase and decrease in a typical day. n training and patron assistance—for a vast array of services such as databases, online searching, e-government (e.g., completing government forms and seeking government information), and others. training can take place formally through classes, but also through point-of-use tutorials requested by patrons. to some extent, librarians commented that, while they do have issues with the public-access computers themselves from time to time, the real challenges that they face regard the actual management of the public- access environment—sign-ups, time limits, cost recovery for print jobs, helping patrons, and so on. one librarian commented that “the computers themselves are pretty stable. we don’t really have too many issues with them per se. it’s everything that goes into, out from, or around the computer that creates issues for us.” as a result of the management challenges, several librar- ies have adopted turn-key solutions, such as public-access management systems (e.g., comprise technology’s smart access manager [http://www.comprisetechnologies .com/product_ .html]) and all-encompassing public computing management systems that include networking and desktops (e.g., userful’s discoverstations [http:// userful.com/libraries/]). these systems allow for an all- in-one sign-up, print cost recovery, filtering (if desired), and security approach. also, the discoverstations are a linux-based, all encompassing public-access management environment. a clear advantage to the discoverstation approach is that the discoverstation is connected to the internet and is accessible by userful staff remotely to update software and perform other maintenance func- tions. they also use open-source operating and applica- tion software. while these solutions do provide efficiencies, they also can create limitations. for example, the discoverstations are a thin-client system and are dependent on the server for graphics and memory, thus limiting their ability to access gaming and social-networking sites. the smart access manager, and similar programs, can rely on smart cards or other technology that users must purchase to print. another limitation is that the time limits are fixed, and, while users get warnings as time runs out, the ses- sion can end abruptly. these approaches are by and large adopted by librar- ies to ease the management associated with public-access computers and let staff concentrate on other duties and responsibilities. one librarian indicated that “until we had our management system, we would spend most of the day signing people up for the computers, or asking them to finish their work for the next person in line.” n planning for pat services and resources public libraries face a number of challenges when plan- ning for pat services and resources. this is primarily because pat planning involves more than computers. any planning needs to encompass n building needs, requirements, limitations, and design; n technology assessment that considers the library’s existing technology, technology potential, current practices, and future trends; n planning for and supporting multiple technology platforms; n telecommunications and networking; n services and resources available in the market- place—those specifically for libraries and those more broadly available to consumers and used by patrons; n specific needs and requirements of technology (e.g., memory, disk space, training, other); n requirements of other it groups with which the library may need to integrate, for example, city or county technology mandates; n support needs, including the need to enter into maintenance agreements for computer, network, and other equipment and software; n staff capabilities, such as current staff skill sets and their ability to handle the technologies under review or purchased; and n policy, such as requirements to filter because of local, state or federal mandates. the above list may not be exhaustive, but rather based on the main items that librarians identified during the site visits, and they serve to provide indicators of the chal- lenges those planning library it initiatives face. information technology and libraries | june n the endless upgrade and planning one librarian likened the pat environment to “being a gerbil on a treadmill. you go round and round and never really arrive,” a reference to the fact that public libraries are in a perpetual cycle of planning and implementing vari- ous pat services and resources. either hardware needs to be updated or replaced, or there is a software update that needs to be installed, or libraries are looking to the next technology coming down the road. in short, the technol- ogy planning to implementation cycle is perpetual. the upgrade and replacement cycle is further exac- erbated by the funding situation in which most public libraries find themselves. increasingly, public library local and state funding, which combined can account for more than percent of library funding, is flat or declining. the most recent series of public library internet studies indicates an increase in reliance by public libraries on fees and fines, fundraising, private foundation, and grant funding to finance collections and technology within libraries. this places key aspects of library operations in the realm of unreliable and one-time funding sources, thus making it difficult for libraries to develop multiyear plans for pat. n multiple support models to cope with pat management and maintenance issues, public libraries are developing various support strategies. the site visits found a number of technology-support approaches in effect, ranging from no it support to highly centralized statewide approaches. the following list describes the technology-support models encoun- tered during the site visits: . no technology support. libraries in this group have neither technology-support staff nor any type of organized technology-support mechanism with existing library staff. nor do they have access to external support providers such as county or city it staff. libraries in this group might rely on vol- unteers or engage in ad hoc maintenance, but by and large have no formal approach to supporting or maintaining their technology. . internal library support without technology staff. in this model, the library provides its own technology support but does not necessarily have dedicated technology staff. rather, the library has desig- nated one or more staff members to serve as the it person. usually this person has an interest in technology but has other primary responsibilities within the library. there may be some structure to the support—such as updating software (e.g., windows patches) once a week at a certain time— but it may be more ad hoc in approach. also, the library may try to provide its designated it person(s) with training to develop his or her skills further over time. . internal library support with technology staff. in this model, the library has at least one dedicated it staff person (part- or full-time) who is responsible for maintaining and planning the library’s pat environment. the person may also have respon- sibilities for network maintenance and a range of technology-based services and resources. at the higher end of this approach are libraries with mul- tiple it staff with differing responsibilities, such as networking, telecommunications, public-access computers, the ils, etc. libraries at this end of the spectrum tend to have a high degree of technology sophistication but may face other challenges (i.e., staffing shortages in key areas). . library consortia. over the years, public libraries have developed consortia for a range of services— shared ilss, resource sharing, resource licensing, and more. as public-library needs evolve, so too do the roles of library consortia. consortia increas- ingly provide training and technology-support services, and may be funded through membership fees, state aid, or other sources. . technology partners. while some libraries may rely on consortia for their technology support, others are seeking libraries that have more technology expertise, infrastructure, and abilities with whom to partner. this can be a fee-for-service arrange- ment that may involve sharing an ils, a mainte- nance agreement for network and public-access computer support, and a range of services. these arrangements allow the partner libraries to have some input into the technology planning and implementation processes without incurring the full expense of testing the technologies, having to implement them first, or hiring necessary staff (e.g., to manage the ils). the disadvantage to this model is that the smaller partner libraries are dependent on the technology decisions that the primary partner makes, including upgrade cycles, technology choices, migration time frames, etc. . city, county, or other agency it support. as city or county government agencies, some libraries receive technology support from the city or county it department (or in some cases the education department). this support ranges from a full slate of services and support available to the library to support only for the staff network and computers. public access technologies in public libraries | bertot even at the higher end of the support spectrum, librarians gave mixed reviews for the support received from it agencies. this was primarily because of competing philosophies regarding the pat environment, with public librarians wanting an open-access policy to allow users access to a range of information service and resources and it agency staff wanting to essentially lock down the public-access environment and thus severely limit the functionality of the public-access computers and network services (i.e., wireless). other limita- tions might include prescribed pat, specified ven- dors, and bidding requirements. . state library support. one state library visited pro- vides a high degree of service through its statewide approach to supporting public-access computing in the state’s public libraries. the state library has it staff in five locations throughout the state to provide support on a regional level but also has additional staff in the capital. these staff offer training, in- house technical support, phone support, and can remote access the public-access computers in public libraries to troubleshoot, update, and perform other functions. moreover, this state built a statewide network through a statewide application to the fed- eral e-rate program, thus providing broadband to all libraries. this model extends the availability of qualified technical support staff to all public librar- ies in the state—by phone as well as in person if need be. as a result, this enables public libraries to concentrate on service delivery to patrons. it is important to note that there are combinations of the above models in public libraries. for example, some libraries support their public-access networks and tech- nology while the county or city it department supports the staff network and technology. it is clear, however, that there are a number of models for technology support in public libraries, and likely more than are presented in this article. the key issue is that public libraries are engaging in a broad spectrum of strategies to support, maintain, and manage their pat infrastructure. also of significance is that there are public libraries that have no technology-support services that provide pat services and resources. these libraries tend to serve populations of less than ten thousand, are rural, have fewer than five full-time equivalents (ftes), and are unlikely to be staffed by professional librarians. staff needs and pressures the study found a number of issues related to the effect of pat on library staff. this section of the findings discusses the primary factors affecting library staff as they work in the public-access context. n multiple skills needed not only is the pace of technological change increasing, but the change requires an ever-increasing array of skills because of the complexity of applications, technolo- gies, and services. an example of such complexity is the library opac or ils. visited libraries indicated that such systems are becoming so complex and technologically sophisticated that there is a need for a full-time staff per- son to run and maintain the library ils. given the range of hardware, software, and network- ing infrastructure, as well as planning and pat manage- ment requirements, public librarians need a number of skills to successfully implement and maintain their pat environments. moreover, the skill needs depend on the librarian’s position—for example, an actual it staff person versus a reference librarian who does double duty by serv- ing as the library’s it person. the skills required fall into technology, information literacy, service and facilities plan- ning, management, and leadership and advocacy areas: n technology o general computer troubleshooting o basic maintenance, such as mouse and key- board cleaning o basic computer repair, such as memory replacement, floppy drive replacement, disk defragmentation, etc. o basic networking, such as troubleshooting an “internet” issue versus a computer problem o telecommunications so as to understand the design and maintenance of broadband net- works o integrated library systems o web design n information literacy o searching and using internet-based resources o searching and using library licensed resources o training patrons on the use of the public- access computers, general internet resources, and library resources o designing curriculum for various patron training courses n services and facilities planning o technology plan development and imple- mentation (including budgeting) o telecommunications planning (including information technology and libraries | june e-rate plan and application development) o building design so as to accommodate the requirements of public access technologies n management o license and contract negotiation for licensed resources, various public-access software and licenses, and maintenance agreements (service and repair agreements) o integration of pat into library operations o troubleshooting guidelines and process o policy development, such as acceptable use, filtering, filtering removal requests by patrons, etc. n leadership and advocacy o grant writing and partnership development so as to fund pat services and resources and extend out into the community that the library serves o advocacy so as to be able to demonstrate the value of pat in the library as a community good o leadership so as to build a community approach to public access with the library as one of the foundational institutions these items provide a broad cross section of the skills that public library staff may need to offer a robust pat environment. in the case of smaller, rural libraries, these requirements in general fall to the library director—along with all other duties of running the public library. in librar- ies that have separate technology, collections development, and other specialized staff, the skills and expertise may be dispersed throughout various areas in the library. n training public librarians receive a range of technology training— including none at all. in some cases, this might be a basic workshop on some aspect of technology at a state library association annual meeting or a regional workshop hosted by the library’s consortium. it could be an online course through webjunction (http://www.webjunction .org/). it could be a one-on-one session with a vendor representative or colleague. or it could be a formal, mul- tiday class regarding the latest release of an ils. if avail- able, public librarians have access to technology training that can take many forms, has a wide array of content (basic to expert), and can enhance staff knowledge about it with varying degrees of success. an issue raised by librarians was that having access to training and being able to take advantage of training are two separate things. regardless of the training delivery medium, librarians indicated that they were not always able to get release time to attend a training session. this was particularly the case for small, rural libraries that had less than five ftes spread out over several part-time individuals. for these staff to take advantage of train- ing would require a substitute to cover public-service hours—or shut down the library. funding information technology as one might expect, there was a range of technology budgets in the public libraries visited or interviewed— from no technology budget to a substantial technology budget, and many points in between. some libraries had a dedicated it budget line item, others had only an oper- ating budget out of which they might carve some funds for technology. libraries with dedicated it budgets by and large had at least one it staff person; libraries with no it budget largely relied on a staff person responsible for other library functions to manage their technology. in the smallest libraries, the library director served as the tech- nology specialist in addition to being the general library operation manager. some libraries have established foundations through which they can raise funds for technology, among other library needs. many seek grants and thus devote substan- tial effort to seeking grant initiatives and writing grant proposals. some libraries held fundraisers and worked with their library friends groups to generate funds. other libraries engage in all of the above efforts to provide for their pat infrastructure, services, and resources. in short, there are several budgetary approaches public libraries use to support their pat environment. critical to note is that a number of libraries are increasingly relying on nonrecur- ring funds to support pats, a fact corroborated by the and public library internet surveys. the buildings when one visits public libraries, one is immediately struck by the diversity in design, functionality, and archi- tecture of the buildings. public libraries often reflect the communities that they serve not only in the collection and service, but also in the facilities. this diversity serves the public library community well because it allows for a custom approach to libraries and their community. the building design, however, can also be a source of substantial challenge for public libraries. the increased integration of technology into library service places a range of stresses on buildings—physical space for work- stations and other equipment and specialized furniture, power, server rooms, and cabling, for example. along with the library-based technology requirements come those of patrons—particularly the need for power so that public access technologies in public libraries | bertot patrons may plug in their laptops or other devices. also important to note is that the building limitations also extend to staff and their access to computing and net- worked technologies. a number of librarians commented that they are “sim- ply at capacity.” one librarian summed it up by stating that “there’s no more room at the inn. unless we start removing parts of our collection, we don’t have any more room for workstations.” another said that, “while we do have the space to add more computers, we don’t have enough power or outlets to support them. and, with our building, it’s not a simple thing to add.” in short, many libraries are reaching, or have reached, a saturation point as to just how much pat they can support. n discussion and implications over time, pat services have become essential services that public libraries provide their communities. with nearly all public libraries connected to the internet and offering public-access computers, the high percentage of libraries that offer internet-based services and resources, the overall usage of these resources by the public, and percent of public libraries reporting that they are the only free provider of pat in their communities, it is clear that the provision of pat services is a key and critical ser- vice role that public libraries offer. it is also clear, how- ever, that the extent to which public libraries can continue to absorb, update, and expand their pat depends on the resolution of a number of staffing, financial, maintenance and management, and building barriers. in a time of constrained budgets, it is unlikely that libraries will receive increased operational funding. indeed, reports of library funding cuts are increasing in the current economic downturn, which affects the ability of libraries to increase, or significantly update, staff—particularly in the areas of technology, licensing additional resources, procuring additional and new com- puters, and purchasing and offering expanded services such as digital photography, gaming, or social network- ing. moreover, the same financial constraints can affect the ability of libraries to raise capital funds for building improvements and new construction. funding also has an effect on the training that public libraries can offer or develop for their staff. and training is becoming increasingly important to the success of pat services and resources in public libraries—but not just training regarding the latest technologies. rather, there is a need for training that provides instruction on the rela- tionship between the level of pat services and resources a library can or desires to provide and advocacy; broad- band, computing, and other needs; technology planning and management; collaboration and partnering; and leadership. the public library pat environment is com- plex, encompasses a number of technologies, and has ties to many community services and resources. training programs need to reflect this complexity. the continued provision of pat services in public libraries is increasingly burdensome on the public library community, and the pressures to expand their pat ser- vices and resources continues to grow—particularly as libraries report their “sole provider” of free pat status in their communities. the successful libraries in terms of pat services and resources visited had staff that could n understand pat (both in terms of functionality and potential); n think creatively across the technology and library service spectrum; n integrate online content, pat, and library services; n articulate the value of pat as an essential commu- nity need and public library service; n articulate the role of the perception of the library by its community as a critical bridge to online con- tent; n demonstrate leadership within the community and library; n form partnerships and extend pat services and resources into the community; and n raise funds and develop other support mecha- nisms to enhance pat services and resources in the library and throughout the community. in short, successful pat in libraries was being rede- fined in the context of communitywide pat service and resource provision. this approach not only can lead to a more robust community pat infrastructure, but it also lessens the library’s burden of pat service and resource provision. but equally important to note is that the extent to which all public libraries can engage in these activities on their own is unclear. indeed, several libraries visited were struggling to maintain basic pat service levels and indi- cated that increasing pat services came at the expense of other library services. “we’re trying to meet demand,” one librarian said, “but we have too few computers, too slow a connection, and staff don’t always know what to do when things go wrong or someone comes in talking about the latest technology or website.” for some librar- ies, therefore, quality pat services that meet community needs are simply out of reach. thus another implication and finding of the study is the need for libraries to explore other models of sup- port for their pat environments—for example, using the services of a regional cooperative, if available; if none is available, libraries could form their own cooperative for resource sharing, technology support, and other aspects of pat service provision. the same approach could be information technology and libraries | june taken within a city or county to enhance technology sup- port throughout a region. another approach would be to outsource a library’s pat support and maintenance to a nearby library with support staff in a fee-for-service approach. there are a number of approaches that librar- ies could take to support their pat infrastructure. a key point is that libraries need to consider pat service provi- sion in a broader community, regional, or state context, and the study found some libraries doing so. the need to avail staff of the skills required to truly support pat was a recurring theme throughout the site visits. approaches and access to training var- ied. for example, some state libraries provided—either directly or through the hiring of consultants and instruc- tors—a number of technology-related courses taught in regional locations. an example of this approach is california’s infopeople project (http://www.infopeople .org/). some state libraries subscribed to webjunction (http://www.webjunction.org/), which provides access to online instructional content. online manuals provided by compumentor through a grant funded by the bill and melinda gates foundation aimed at helping rural libraries support their pat (www.maintainitproject.org) are another resource. beyond technology skills training, however, is the need for technology planning, effective communica- tion, leadership, value demonstration, and advocacy. the extent to which leadership, advocacy, and library market- ing, for example, are able to be taught remains a question. all of these issues take place with the backdrop of an economic downturn and budgetary constraints. increased operating costs created through inflation and higher energy costs place substantial pressures on public libraries simply to maintain current levels of service— much less engage in the additional levels of service that the pat environment brings. indeed, as the public library funding and technology access study demonstrated, public libraries are increasingly funding their technology-based services through non-recurring funds such as fines and fundraising activities. thus, the ability of public libraries to provide robust pat services and resources is increasingly limited unless such service provision comes at the expense of other library services. alone, the financial pressures place a high burden on public libraries. combined with the building, staffing, skills, and other constraints reported by public libraries, however, the emerging picture for library pat services and resources is one of significant challenge. n three key areas for additional exploration the findings from the study point to the need for addi- tional research and exploration of three key services areas and issues related to pat support and services: . develop a better understanding of success in the pat environment. this study and the study by bertot et al. point to what is required for libraries to be successful in a networked environment. in fact, the public libraries and the internet report contained a section entitled “the successfully networked public library,” which offered a range of checklists for public libraries (and others) to consider as they planned and implemented their networked services. this study identified addi- tional success factors and considerations focused specifically on the public access technology envi- ronment. together, these efforts point to the need to better understand and articulate the critical suc- cess factors necessary for public libraries to plan, implement, and update their pat given current service contexts. this is particularly necessary in the context of meeting user expectations and needs regarding networked technologies and services. . further identify technology-support models. this study uncovered a number of different technology- support models implemented by public libraries. undoubtedly there are additional models that require identification. but, more importantly, there is a need to further explore how each technology- support model assists libraries, under what cir- cumstances, and in what ways. some models may be more or less appropriate on the basis of the ser- vice context of the library—and that is not clearly understood at this time. . levels of service capabilities. an underlying theme throughout this research, and one that is increasingly supported by the public library and the internet studies, is that the pat service context is essentially a continuum from low service and capability to high service and capability. there are a number of factors contributing to where libraries may lie on the success continuum—funding, management, leadership, attitude, skills, community support, and innovation, to name a few. this continuum requires additional research, and the research implications could be profound. emerging data indicate that there are public libraries that will be unable to con- tinue to evolve and meet the increased demands of the networked environment, both in terms of staff and infrastructure. public libraries will have to make choices regarding the provision of pat ser- vices and resources in light of their ability to provide high-quality services (as defined by their service communities). for better or worse, the technology environment continually evolves and requires new technologies, management, and support. that is, public access technologies in public libraries | bertot and will continue to be, the nature of public access to the internet. though there are likely other issues worthy of explo- ration, these three are critical to further our understand- ing of the pat environment and public library roles and issues associated with the provision of public access. n conclusion the pat environment in which public libraries operate is increasingly complex and continues to grow in funding, maintenance and management, staffing, and building demands. public libraries have navigated this environ- ment successfully for more than fifteen years; however, stresses are now evident. libraries rose quickly to the challenge of providing public-access services to the com- munities that they serve. the challenges libraries face are not necessarily insurmountable, and there are a range of tools designed to help public libraries plan and man- age their public-access services. these tools, however, place the burden of public access, or assume that the burden of public access in placed, on the public library. given increased operating costs because of inflation, the continual need to innovate and upgrade technologies, staff technology skills requirements, and other factors discussed in this article, libraries may not be in a position to shoulder the burden of public access alone. thus there is a need to reconsider the extent to which pat provision is the sole responsibility of the library; perhaps there is a need to integrate and expand public access throughout a community. the potential of such an approach can benefit a community through an integrated and broader access strategy, but also can relieve the pressure on the public library as the sole provider of public access. n acknowledgement this reserach was made possible in part through the sup- port of the maintianit project (http://www.maintainit project.org/), an effort of the nonprofit techsoup web resource (http://www.techsoup.org/). references . charles r. mcclure, john carlo bertot, and douglas l. zweizig, public libraries and the internet: study results, policy issues, and recommendations (washington, d.c.: national commission on libraries and information science, ). . john carlo bertot and charles r. mcclure, moving toward more effective public internet access: the national survey of public library outlet internet connectivity (washington, d.c.: national commission on libraries and information science, ), http://www.liicenter.org/reports/ _plinternet_ study.pdf (accessed apr. , ). . charles r. mcclure, john carlo bertot, and john c. beachboard, internet costs and cost models for public libraries (washington, d.c.: national commission on libraries and information science, ). . charles r. mcclure, john carlo bertot, and douglas l. zweizig, public libraries and the internet: study results, policy issues, and recommendations (washington, d.c.: national com- mission on libraries and information science, ); john carlo bertot, charles r. mcclure, paul t. jaeger, and joe ryan, public libraries and the internet : study results and findings (tallahassee, fla.: information institute, ), http://www .ii.fsu.edu/projectfiles/plinternet/ / _plinternet.pdf (accessed mar. , ). . john carlo bertot, charles r. mcclure, carla b. wright, elise jensen, and susan thomas, public libraries and the internet : study results and findings (tallahassee, fla.: information institute, ). http://www.ii.fsu.edu/projectfiles/plinternet/ / _plinternet.pdf (accessed sept. , ). . charles r. mcclure and paul t. jaeger, public libraries and internet service roles: measuring and maximizing internet services (chicago: ala, ). . george d’elia, june abbas, kay bishop, donald jacobs, and eleanor jo rodger, “the impact of youth’s use of the inter- net on the use of the public library,” journal of the american soci- ety for information science & technology , no. ( ): – ; george d’elia, corinne jorgensen, joseph woelfel, and eleanor jo rodger, “the impact of the internet on public library use: an analysis of the current consumer market for library and internet services,” journal of the american society for information science & technology , no. ( ): – . . national center for education statistics (nces), public libraries in the united states: fiscal year [nces ] (washington, d.c.: national center for education statistics, ); pew american and internet life, “internet activities,” http:// www.pewinternet.org/trends/internet_activities_ . . .htm (accessed mar. , ). . bertot et al., public libraries and the internet . . ibid. . cheryl bryan, managing facilities for results: optimiz- ing space for services (chicago: public library association, ); joseph matthews, strategic planning and management for library managers (westport, conn.: libraries unlimited, ); joseph matthews, technology planning: preparing and updating a library technology plan (westport, conn.: libraries unlimited, ); diane mayo and jeanne goodrich, staffing for results: a guide to working smarter (chicago: public library association, ). . ala, libraries connect communities: public library fund- ing & technology access study (chicago: ala, ), http:// www.ala.org/ala/aboutala/offices/ors/plftas/ report.cfm (accessed mar. , ). . charles p. smith, ed., motivation and personality: hand- book of thematic content analysis (new york: cambridge univ. information technology and libraries | june pr., ); klaus krippendorf, content analysis: an introduction to its methodol- ogy (beverly hills, calif.: sage, ). . ala, libraries connect communities. . bertot et al., public libraries and the internet ; bertot et al., public libraries and the internet . . ibid. . nces, public libraries in the united states. . bertot et al., public libraries and the internet . . american libraries, “branch clos- ings and budget cuts threaten libraries nationwide,” nov. , , http://www .ala.org/ala/alonline/currentnews/ newsarchive/ /november / branchesthreatened.cfm (accessed nov. , ). . ala, libraries connect communities. . bertot et al., public libraries and the internet . . bertot et al., public libraries and the internet . information technology and libraries | march mary kurtz dublin core, dspace, and a brief analysis of three university repositories this paper provides an overview of dublin core (dc) and dspace together with an examination of the institutional repositories of three public research universities. the uni- versities all use dc and dspace to create and manage their repositories. i drew a sampling of records from each reposi- tory and examined them for metadata quality using the criteria of completeness, accuracy, and consistency. i also examined the quality of records with reference to the meth- ods of educating repository users. one repository used librarians to oversee the archiving process, while the other two employed two different strategies as part of the self- archiving process. the librarian-overseen archive had the most complete and accurate records for dspace entries. t he last quarter of the twentieth century has seen the birth, evolution, and explosive proliferation of a bewildering variety of new data types and formats. digital text and images, audio and video files, spreadsheets, websites, interactive databases, rss feeds, streaming live video, computer programs, and macros are merely a few examples of the kinds of data that can be now found on the web and elsewhere. these new dataforms do not always conform to conventional cata- loging formats. in an attempt to bring some sort of order from chaos, the concept of metadata (literally “data about data”) arose. metadata is, according to ala, “structured, encoded data that describe characteristics of information- bearing entities to aid in the identification, discovery, assessment, and management of the described entities.” metadata is an attempt to capture the contextual information surrounding a datum. the enriching con- textual information assists the data user to understand how to use the original datum. metadata also attempts to bridge the semantic gap between machine users of data and human users of the same data. n dublin core dublin core (dc) is a metadata schema that arose from an invitational workshop sponsored by the online computer library center (oclc) in . “dublin” refers to the location of this original meeting in dublin, ohio, and “core” refers to that fact dc is set of metadata elements that are basic, but expandable. dc draws upon concepts from many disciplines, including librarianship, computer science, and archival preservation. the standards and definitions of the dc element sets have been developed and refined by the dublin core metadata initiative (dcmi) with an eye to interoperabil- ity. dcmi maintains a website (http://dublincore.org/ documents/dces/) that hosts the current definitions of all the dc elements and their properties. dc is a set of fifteen basic elements plus three addi- tional elements. all elements are both optional and repeatable. the basic dc elements are: . title . creator . subject . description . publisher . contributor . date . type . format . identifier . source . language . relation . coverage . rights the additional dc elements are: . audience . provenance . rights holder dc allows for element refinements (or subfields) that narrow the meaning of an element, making it more specific. the use of these refinements is not required. dc also allows for the addition of nonstandard elements for local use. n dspace dspace is an open-source software package that provides management tools for digital assets. it is frequently used to create and manage institutional repositories. first released in , dspace is a joint development effort of hewlett packard (hp) labs and the massachusetts institute of technology (mit). today, dspace’s future mary kurtz (mhkurtz@gmail.com) is a june graduate of drexel university’s school of information technology. she also holds a bs in secondary education from the university of scran- ton and an ma in english from the university of illinois at urbana– champaign. currently, kurtz volunteers her time in technical ser- vices/cataloging at simms library at albuquerque academy and in corporate archives at lovelace respiratory research institute (www.lrri.org), where she is using dspace to manage a diverse collection of historical photographs and scientific publications. dc, dspace, and a brief analysis of three university repositories | kurtz is guided by a loose grouping of interested developers called the dspace committers group, whose members currently include hp labs, mit, oclc, the university of cambridge, the university of edinburgh, the australian national university, and texas a&m university. dspace version . was released in and the newest version, dspace . , was released in march . more than one thousand institutions around the world use dspace, including public and private colleges and universities and a variety not-for-profit corporations. dc is at the heart of dspace. although dspace can be customized to a limited extent, the basic and quali- fied elements of dc and their refinements form dspace’s backbone. n how dspace works: a contributor’s perspective dspace is designed for use by “metadata naive” contribu- tors. this is a conscious design choice made by its devel- opers and in keeping with the philosophy of inclusion for institutional repositories. dspace was developed for use by a wide variety of contributors with a wide range of metadata and bibliographic skills. dspace simplifies the metadata markup process by using terminology that is different from dc standards and by automating the production of element fields and xml/html code. dspace has four hierarchical levels of users: users, contributors, community administrators, and network/ systems administrators. the user is a member of the general public who will retrieve information from the repository via browsing the database or conducting structured searches for specific information. the contributor is an individual who wishes to add their own work to the database. to become a contributor, one must be approved by a dspace community adminis- trator and receive a password. a contributor may create, upload, and (depending upon the privileges bestowed upon him by his community administrator), edit or remove informational records. their editing and removal privileges are restricted to their own records. a community administrator has oversight within their specialized area of dspace and accordingly has more privileges within the system than a contributor. a community administrator may create, upload, edit, and remove records, but also can edit and remove all records available within the community’s area of the database. additionally, the community administrator has access to some metadata about the repository’s records that is not available to users and contributors and has the power to approve requests to become contributors and grant upload access to the database. lastly, the commu- nity administrator sets the rights policy for all materials included in the database and writes the statement of rights that every contributor must agree to with every record upload. the network/systems administrator is not involved with database content, focusing rather on software main- tenance and code customization. when a dspace contributor wishes to create a new record, the software walks them through the process. dspace presents seven screens in sequence that ask for specific information to be entered via check buttons, fill- in textboxes, and sliders. at the end of this process, the contributor must electronically sign an acceptance of the statement of rights. because dspace’s software attempts to simplify the metadata-creation process for contributors, its terminol- ogy is different from dc’s. dspace uses more common terms that are familiar to a wider variety of individu- als. for example, dspace asks the contributor to list an “author” for the work, not a “creator” or a “contribu- tor.” in fact, those terms appear nowhere in any dspace. instead, dspace takes the text entered in the author textbox and maps it to a dc element—something that has profound implications if the mapping does not follow expected dc definitions. likewise, dspace does not use “subject” when asking the contributor to describe their material. instead, dspace asks the contributor to list keywords. text entered into the keyword field is then mapped into the subject ele- ment. while this seems like a reasonable path, it does have some interesting implications for how the subject element is interpreted and used by contributors. dc’s metadata elements are all optional. this is not true in dspace. dspace has both mandatory and auto- matic elements in its records. because of this, data records created in dspace look different than data records created in dc. these mandatory, automatic, and default fields affect the fill frequency of certain dc elements—with all of these elements having percent participation. in dspace, the title element is mandatory; that is, it is a required element. the software will not allow the contributor to proceed if the title text box is left empty. as a consequence, all dspace records will have percent participation in the title element. dspace has seven automatic elements, that is, ele- ment fields that are created by the software without any need for contributor input. three are date elements, two are format elements, one is an identifier, and one is provenance. dspace automatically records the time of the each record’s creation in machine-readable form. when the record is uploaded into the database, this time- stamp is entered into three element fields: dc.date.avail- able, dc.date.accessioned, and dc.date.issued. therefore dspace records have percent participation in the date element. for previously published materials, a separate screen asks for the original publication date, which is then information technology and libraries | march placed in the dc.date.issued element. like title, the origi- nal date of publication is a mandatory field, and failure to enter a meaningful numerical date into the textbox will halt the creation of a record. in a similar manner, dspace “reads” the kind of file the contributor is uploading to the database. dspace automatically records the size and type (.doc, .jpg, .pdf, etc.) of the file or files. this data is automatically entered into dc.format.mimetype and dc.format.extent. like date, all dspace records will have percent participation in the format element. likewise, dspace automatically assigns a location identifier when a record is uploaded to the database. this information is recorded as an uri and placed in the identifier element. all dspace records have a dc.identifier.uri field. the final automatic element is provenance. at the time of record creation, dspace records the identity of the contributor (derived from the sign-in identity and pass- word) and places this information into a dc.provenance element field. this information becomes a permanent part of the dspace record; however, this field is a hidden to users. typically only community and network/sys- tems administrators may view provenance information. still, like date, format, and identifier elements, dspace records have automatic percent participation in prov- enance. because of the design of dspace’s software, all dspace-created records will have a combination of both contributor-created and dspace-created metadata. all dspace records can be edited. during record cre- ation, the contributor may at any time move backward through his record to alter information. once the record has been finished and the statement of rights signed, the completed record moves into the community administra- tor’s workflow. once the record has entered the workflow, the community administrator is able to view the record with all the metadata tags attached and make changes using dspace’s editing tools. however, depending on the local practices and the volume of records passing through the administrator’s workflow, the administrator may simply upload records without first reviewing them. a record may also be edited after it has been uploaded, with any changes being uploaded into the database at the end of editing process. in editing a record after it has been uploaded, the contributor, providing he has been granted the appropriate privileges, is able to see all the metadata elements that have attached to the record. calling up the editing tools at this point allows the contributor or admin- istrator to make significant changes to the elements and their qualifiers, something that is not possible during the record’s creation. when using the editing tools, the simpli- fied contributor interface disappears, and the metadata elements fields are labeled with their dc names. the con- tributor or administrator may remove metadata tags and the information they contain and add new ones selecting the appropriate metadata element and qualifier from a slider. for example, during the editing process, the contrib- utor or administrator may choose to create dc.contributor. editor or dc.subject.lcsh options—something not possible during the record-creation process. in the examination of the dspace records from our three repositories, dspace’s shaping influence on element participation and metadata quality will be clearly seen. n the repositories dspace is principally used by academic and corporate nonprofit agencies to create and manage their insti- tutional repositories. for this study, i selected three academic institutions that shared similar characteristics (large, public, research-based universities) but which had differing approaches to how they managed their metadata-quality issues. the university of new mexico (unm) dspace reposi- tory (dspaceunm) holds a wide-ranging set of records, including materials from the university’s faculty and administration, the law school, the anderson school of business administration, and the medical school, as well as materials from a number of tangentially related university entities like the western water policy review advisory commission, new mexico water trust board, and governor richardson’s task force on ethic reform. at the time of the initial research for this paper (spring ), dspaceunm provided little easily acces- sible on-site education for contributors about the dspace record-creation process. what was offered—a set of eight general information files—was buried deep inside the library community. a contributor would have to know the files existed to find them. by summer , this had changed. dspaceunm had a new homepage layout. there is now a link to “help sheets and promotional materials” at the top center of the homepage. this link leads to the previously difficult-to- find help files. the content of the help files, however, remains largely unchanged. they discuss community creation, copy- rights, administrative workflow for community creation, a list of supported formats, a statement of dspaceunm’s privacy policy, and a list of required, encouraged, and not required elements for each new record created. for the most part, dspaceunm help sheets do not attempt to educate the contributor in issues of metadata quality. there is no discussion of dc terminology, no attempts to refer the contributor to a thesaurus or controlled vocabu- lary list, nor any explanation of the record-creation or editing process. this lack of contributor education may be explained in part because dspaceunm requires all new records dc, dspace, and a brief analysis of three university repositories | kurtz to be reviewed by a subject area librarian as part of the dspace community workflow. thus any contributor errors, in theory, ought to be caught and corrected before being uploaded to the database. the university of washington (uw) dspace reposi- tory (researchworks at the university of washington) hosts a narrower set of records than dspaceunm, with the materials limited to the those contributed by the university’s faculty, students, and staff, plus materials from the uw’s archives and uw’s school of public and community health. in , researchworks was self-archiving. most contributors were expected to use dspace to create and upload their record. there is no indication in the publicly available information about the record creation workflow if record reviews were conducted before record upload. the help link on the researchworks homepage brought contributors to a set of screen-by-screen instructions on how to use dspace’s software to create and upload a record. the step-through did not include instructions on how to edit a record once it had been created. no expla- nation of the meanings or definitions of the various dc elements was included in the help files. there also were no suggestions about the use of a controlled vocabulary or a thesaurus for subject headings. by , this link had disappeared and the associated contributor education materials with it. the knowledge bank at ohio state university(osu) is the third repository examined for this paper. osu’s repository hosts more than thirty communities, all of which are associated with various academic departments or special university programs. like researchworks at uw, osu’s repository appears to be self-archiving with no clear policy statement as to whether a record is reviewed before it is uploaded to the repository’s database. osu makes a strong effort to educate its contribu- tors. on the upper-left of the knowledge bank homepage is a slider link that brings the contributor (or any user) to several important and useful sources of repository information: about knowledge bank, faqs, policies, video upload procedures, community set-up form, describing your resources, and knowledge bank licensing agreement. the existence and use of metadata in knowledge bank are explicitly mentioned in the faq and policies areas, together with an explanation of what metadata is and how metadata is used (faq), and a list of sup- ported metadata elements (policies). the describe your resources section gives extended definitions of each dspace-available dc metadata element and provides examples of appropriate metadata-element use. knowledge bank provides the most comprehensive contributor education information of any of the three repositories examined. it does not use a controlled vocabulary list for subject headings, and it does not offer a thesaurus. n data and analysis i chose twenty randomly selected full records from each repository. no more than one record was taken from any one collection to gather a broad sampling from each repository. i examined each record for the quality of its metadata. metadata quality is a semantically slippery term. park, in the spring special metadata issue of cataloging and classification quarterly, suggested that most com- monly accepted criteria for metadata quality are com- pleteness, accuracy, and consistence. those criteria will be applied in this analysis. for the purpose of this paper, i define completeness as the fill rate for key metadata elements. because the purpose of metadata is to identify the record and to assist in the user’s search process, the key elements are title, contributor/creator, subject, and description.abstract— all contributor-generated fields. i chose these elements because these are the fields that the dspace software uses when someone conducts an unrestricted search. table shows the fill rate for the title element is percent for all three repositories. this is to be expected because, as noted above, title is mandatory field. the fill rate for contributor/creator is likewise high: of ( percent) for unm, of ( percent) for uw, and of ( percent) for osu. (osu’s fill rate for creator and contributor were summed because osu uses different definitions for creator and contributor element fields than do unm or uw. this discrepancy will be discussed in greater depth in the consistency of metadata terminology below.) the fill rate for subject was more variable. unm’s subject fill rate was percent, while uw’s was per- cent, and osu’s was percent. the fill rate for the description.abstract subfield was of ( percent) at unm, of ( percent) at uw, and of ( percent) at osu. (see appendix a for a complete list of metadata elements and subfields used by each of the three repositories.) the relatively low fill rate (below percent) at the osu knowledgebank in both subject and description .abstract suggests a lack of completeness in that reposi- tory’s records. accuracy in metadata quality is the essential “cor- rectness” of a record. correctness issues in a record range from data-entry issues (typos, misspellings, and inconsis- tent date formats) to the correct application of metadata definitions and data overlaps. accuracy is perhaps the most difficult of the metadata information technology and libraries | march quality criteria to judge. local practices vary widely, and dc allows for the creation of custom metadata tags for local use. additionally, there is long-standing debate and confusion about the definitions of metadata elements even among librarians and information professionals. because of this, only the most egregious of accuracy errors were considered for this paper. all three repositories had at least one record that contained one or more inaccurate metadata fields; two of them had four or more inaccurate records. inaccurate records included a wide variety of accu- racy errors, including poor subject information (no matter how loosely one defines a subject heading, “the” is not an accurate descriptor); mutually contradictory metadata (record contained two different language tags, although only one applied to the content); and one in which the abstract was significantly longer and only tangentially related than the file it described. additionally, records showed confusion over contributor versus creator ele- ments. in a few records, contributors entered duplicate information into both element fields. this observation supports park and childress’s findings that there is wide- spread confusion over these elements. among the most problematic records in terms of accuracy were those contained in uw’s early buddhist manuscripts project. this collection, which has been removed from public access since the original data was drawn for this paper, contained numerous ambiguous, contradictory, and inaccurate metadata elements. while contributor-generated subject headings were specifically not examined for this paper, it must be noted that was a wide variation in the level of detail and vocab- ulary used to describe records. no community within any of the repositories had specific rules for the generation of keyword descriptors for records, and the lack of guidance shows. consistency can be defined as the homogeneity of formats, definitions, and use of dc elements within the records. this consistency, or uniformity, of data is impor- tant because it promotes basic semantic interoperability. consistency both inside the repository itself and with other repositories makes the repository easier to use and provides the user with higher quality information. all three repositories showed percent consistency in dspace-generated elements. dspace’s automated cre- ation of date and format fields provided reliably consis- tent records in those element fields. dspace’s automatic formatting of personal names in the dc.contributor.author and dc.creator fields also provided excellent internal con- sistency. however, the metadata elements were much less consistent for contributor-generated information. inconsistency within the subject element is where most problems occurred. personal names used as subject heading and capitalization within subject headings both proved to be particular issues. dspace alphabetizes sub- ject headings according to the first letter of the free text entered in the keyword box. thus the same name entered in different formats (first name first or last name first) generates different subject-heading listings. the same is true for capitalization. any difference in capitalization of any word within the free-text entry generates a separate subject heading. another field where consistency was an issue was dc.description.sponsorship. sponsorship is problem because different communities, even different collections within the same community, use the field to hold differ- ent information. some collections used the sponsorship field to hold the name of a thesis or dissertation advisor. some collections used sponsorship to list the funding agency or underwriter for a project being documented inside the record. some collections used sponsorship to acknowledge the donation of the physical materials docu- mented by the record. while all of these are valid uses of the field, they are not the same thing and do not hold the same meaning for the user. the largest consistency issue, however, came from table . metadata fields and their frequencies element univ. of n.m. univ. of wash. ohio state univ. title creator subject description publisher contributor date type identifier source language relation coverage rights provenance ** ** ** **provenance tags are not visible to public users dc, dspace, and a brief analysis of three university repositories | kurtz a comparison of repository policies regarding element use and definition. unaltered dspace software maps contributor-generated information entered into the author textbox during the record-creation process into the dc.contributor.author field. however, osu’s dspace software has been altered so that the dc.contributor .author field does not exist. instead, text entered into the author textbox during the record-creation process maps to dc.creator. although both uses are correct, this choice does create a significant difference in element definitions. osu’s dspace author fields are no longer congruent with other dspace author fields. n conclusions dspace was created as repository management tool. by streamlining the record creation workflow and partially automating the creation of metadata, dspace’s develop- ers hoped to make institutional repositories more useful and functional while time providing an improved experi- ence for both users and contributors. in this, dspace has been partially successful. dspace has made it easier for the “metadata naive” contributor to create records. and, in some ways, dspace has improved the quality of repository metadata. its automatically generated fields ensure better consistency in those elements and subfields. its mandatory fields guarantee percent fill rates in some elements, and this contributes to an increase in metadata completeness. however, dspace still relies heavily on contributor- generated data to fill most of the dc elements, and it is in these contributor-generated fields that most of the metadata quality issues arise. nonmandatory fields are skipped, leading to incomplete records. data entry errors, a lack of authority control over subject headings, and con- fusion over element definitions can lead to poor metadata accuracy. a lack of enforced, uniform naming and capi- talization conventions leads to metadata inconsistency, as does the localized and individual differences in the application of metadata element definitions. while most of the records examined in this small survey could be characterized as “acceptable” to “good,” some are abysmal. to improve the inconsistency of the dspace records, the three universities have tried differ- ing approaches. only unm’s required record review by a subject area librarian before upload seems to have made any significant impact on metadata quality. unm has a percent fill rate for subject elements in its records, while uw and osu do not. this is not to say that unm’s process is perfect and that poor records do not get into the system—they do (see appendix b for an example). but it appears that for now, the intermediary interven- tion of a librarian during the record-creation process is an improvement over self-archiving—even with educa- tion—by contributors. references and notes . association of library collections & technical services, committee on cataloging: description & access, task force on metadata, “final report,” june , , http://www.libraries .psu.edu/tas/jca/ccda/tf-meta .html (accessed mar. , ). . a voluntary (and therefore less-than-complete) list of current dspace users can be found at http://www.dspace. org/index.php?option=com_content&task=view&id= &ite mid= . further specific information about dspace, includ- ing technical specifications, training materials, licensing, and a user wiki, can be found at http://www.dspace.org/index .php?option=com_content&task=blogcategory&id= &itemi d= . . jung-ran park “metadata quality in digital repositories: a survey of the current state of the art,” cataloging & classifica- tion quarterly , no. ( ): – . . sarah currier et al., “quality assurance for digital learning object repositories: issues for the metadata creation process,” alt-j: research in learning technology , no. ( ): – . . jung-ran park and eric childress, “dc metadata seman- tics: an analysis of the perspectives of informational profession- als,” journal of information science , no. ( ): – . . ibid. . for a fuller discussion of the collection’s problems and challenges in using both dspace and dc, see kathleen for- sythe et al., university of washington ealy buddhist manuscripts project in dspace (paper presented at dc- , seattle, wash., sept. –oct. , ), http://dc .ischool.washington.edu/ archive- / forsythe.pdf (accessed mar. , ). lita cover , cover neal-schuman cover oclc index to advertisers information technology and libraries | march appendix a. a list of the most commonly used qualifiers in each repository university of new mexico dc.date.issued ( ) dc.date.accessioned ( ) dc.date.available ( ) dc.format.mimetype ( ) dc.format.extent ( ) dc.identifier.uri ( ) dc.contributor.author ( )) dc.description.abstract ( ) dc.identifier.citation ( ) dc.description.sponsorship ( ) dc.subject.mesh ( ) dc.contributor.other ( ) dc.description.sponsor ( ) dc.date.created ( ) dc.relation.isbasedon ( ) dc.relation.ispartof ( ) dc.coverage.temporal ( ) dc.coverage.spatial ( ) dc.contributor.other ( ) university of washington dc.date.accessioned ( ) dc.date.available ( ) dc.date.issued ( ) dc.format.mimetype ( ) dc.format.extent ( ) dc. identifier.uri ( ) dc.contributor.author ( ) dc.description.abstract ( ) dc.identifier.citation ( ) dc.identifier.issn ( ) dc.description.sponsorship ( ) dc.contributor.corporateauthor ( ) dc.contributor.illustrator ( ) dc.relation.ispartof ( ) ohio state university dc.date.issued ( ) dc.date.available ( ) dc.date.accessioned ( ) dc.format.mimetype ( ) dc.format.extent ( ) dc.identifier.uri ( ) dc.description.abstract ( ) dc.identifier.citation ( ) dc.subject.lcsh ( ) dc.relation.ispartof ( ) dc.description.sponsorship ( ) dc.identifier.other ( ) dc.contributor.editor ( ) dc.contribtor.advisor ( ) dc.identifier.issn ( ) dc.description.duration ( ) dc.relation.isformatof ( ) dc.description.statementofresponsi- bility ( ) dc.description.tableofcontents ( ) appendix b. sample record dc.identifier.uri http://hdl.handle.net/ / dc.description.abstract president schmidly’s charge for the creation of a north golf course community advisory board. dc.format.extent bytes dc.format.mimetype application/pdf dc.language.iso en_us dc.subject president dc.subject schmidly dc.subject north dc.subject golf dc.subject course dc.subject community dc.subject advisory dc.subject board dc.subject charge dc.title community_advisory_board_charge dc.type other information technology and libraries | june andrew k. pace president’s message: lita forever andrew k. pace (pacea@oclc.org) is lita president / and executive director, networked library services at oclc inc. in dublin, ohio. i was warned when i started my term as lita president that my time at the helm would seem fleeting in retro- spect, and i didn’t believe it. i should have. i suppose most advice of that sort falls on deaf ears—advice to children about growing up, advice to newlyweds, advice to new parents. some things you just have to experience. now i am left with that feeling of having worked very hard while not accomplishing nearly enough. it’s time to buy myself some more time. my predecessor, mark beatty, likes to jokingly intro- duce himself in ala circles as “lita has-been” in refer- ence to his role as lita past-president. i say jokingly because he and i both know it is not true. not only does the past-president continue in an active role on the lita board and executive committee, the past-president has the daunting task of acting as the division’s financial officer. just as mark knows well the nature of this elected (but still volunteer) commitment, so michelle frisque, my successor this july, knows that the hard work started as vice-president/ president-elect has two challenging years ahead. being elected lita president is for all intents and purposes a three-year term with shifting responsibili- ties. add to this the possibility of serving on the board beforehand, and it’s likely that one could serve less time for knocking over a liquor store. i’m joking, of course— there’s nothing punitive about being a lita officer; it’s as rewarding as it is challenging. neither is this intended to be a self-congratulatory screed as my last hurrah in print as lita president. i’ve referred repeatedly to the grassroots success of lita’s board, interest groups, dedicated committees, and engaged volunteers. the flatness of our division is often emulated by others. i thoroughly enjoy engagement with the lita membership, face-to-face and virtual recruitment of new members and volunteers, and group meetings to discuss moving lita forward. i love that lita is fun. fun and enjoyment, coupled with my dedication to the profession that i love, is why i plan to make the most of my time, even as a has-been. all those meetings, all that bureaucracy? well, believe it or not, i like the bureaucracy—process works when you learn to work the process—and all those meetings have actually created some excellent feedback for the lita board. changes in ala, changes in the membership, and changes suggested by committees and interest groups all suggest . . . guess what? change. “change” has been a popular theme these days. i’m in that weird minority of people who does not believe that people don’t like to change. i think if the ideas are good, if the destination is worthwhile, then change is possible and even desirable. i’m always geared up for change, for learning from our mistakes, for asking forgiveness on occasion and for permission even less. this is a long-winded way of saying that i think lita is ready for some change. change to the board, change to the committees and interest groups, and changes to our interactions with lita and ala staff. i think ala and the other divisions are anxious for change as well, and i feel confident that lita and its membership can help, even while we change ourselves. don’t ask me today what the details of these changes are. all i can say is that i will be there for them, help see them through, and will be there on the other side to asses which changes worked and which didn’t. one thing i hope does not change is the passion and dedication of the leaders, volunteers, and members of this great organization. i only hope that our ranks grow, even in times of financial uncertainty. lita provides a valuable network of colleagues and friends—this net- work is always valuable, but it is indispensible in times of difficulty. for many, lita represents a second or third divisional membership, but for networking and collegial support, i think we are second to none. i titled my previous column “lita now.” i think it’s safe for me to say now, “lita forever.” can bibliographic data be put directly onto the semantic web? | yee martha m. yee can bibliographic data be put directly onto the semantic web? this paper is a think piece about the possible future of bib- liographic control; it provides a brief introduction to the semantic web and defines related terms, and it discusses granularity and structure issues and the lack of standards for the efficient display and indexing of bibliographic data. it is also a report on a work in progress—an experi- ment in building a resource description framework (rdf) model of more frbrized cataloging rules than those about to be introduced to the library community (resource description and access) and in creating an rdf data model for the rules. i am now in the process of trying to model my cataloging rules in the form of an rdf model, which can also be inspected at http://myee. bol.ucla.edu/. in the process of doing this, i have discov- ered a number of areas in which i am not sure that rdf is sophisticated enough yet to deal with our data. this article is an attempt to identify some of those areas and explore whether or not the problems i have encountered are soluble—in other words, whether or not our data might be able to live on the semantic web. in this paper, i am focusing on raising the questions about the suitability of rdf to our data that have come up in the course of my work. t his paper is a think piece about the possible future of bibliographic control; as such, it raises more complex questions than it answers. it is also a report on a work in progress—an experiment in build- ing a resource description framework (rdf) model of frbrized descriptive and subject-cataloging rules. here my focus will be on the data model rather than on the frbrized cataloging rules for gathering data to put in the model, although i hope to have more to say about the latter in the future. the intent is not to present you with conclusions but to present some questions about data modeling that have arisen in the course of the experiment. my premise is that decisions about the data model we follow in the future should be made openly and as a com- munity rather than in a small, closed group of insiders. if we are to move toward the creation of metadata that is more interoperable with metadata being created outside our community, as is called for by many in our profes- sion, we will need to address these complex questions as a community following a period of deep thinking, clever experimentation, and astute political strategizing. n the vision the semantic web is still a bewitching midsummer night’s dream. it is the idea that we might be able to replace the existing html–based web consisting of marked-up documents—or pages—with a new rdf– based web consisting of data encoded as classes, class properties, and class relationships (semantic linkages), allowing the web to become a huge shared database. some call this web . , with hyperdata replacing hyper- text. embracing the semantic web might allow us to do a better job of integrating our content and services with the wider internet, thereby satisfying the desire for greater data interoperability that seems to be widespread in our field. it also might free our data from the propri- etary prisons in which it is currently held and allow us to cooperate in developing open-source software to index and display the data in much better ways than we have managed to achieve so far in vendor-developed ils opacs or in giant, bureaucratic bibliographic empires such as oclc worldcat. the semantic web also holds the promise of allow- ing us to make our work more efficient. in this bewitch- ing vision, we would share in the creation of uniform resource identifiers (uris) for works, expressions, mani- festations, persons, corporate bodies, places, subjects, and so on. at the uri would be found all of the data about that entity, including the preferred name and the vari- ant names, but also including much more data about the entity than we currently put into our work (name-title and title), such as personal name, corporate name, geographic, and subject authority records. if any of that data needed to be changed, it would be changed only once, and the change would be immediately accessible to all users, libraries, and library staff by means of links down to local data such as circulation, acquisitions, and binding data. each work would need to be described only once at one uri, each expression would need to be described only once at one uri, and so forth. very much up in the air is the question of what institu- tional structures would support the sharing of the creation of uris for entities on the semantic web. for the data to be reliable, we would need to have a way to ensure that the system would be under the control of people who had been educated about the value of clean and accurate entity definition, the value of choosing “most commonly known” preferred forms (for display in lists of mul- tiple different entities), and the value of providing access martha m. yee (myee@ucla.edu) is cataloging supervisor at the university of california, los angeles film and television archive. information technology and libraries | june under all variant forms likely to be sought. at the same time, we would need a mechanism to ensure that any interested members of the public could contribute to the effort of gathering variants or correcting entity definitions when we have had inadequate information. for example, it would be very valuable to have the input of a textual or descriptive bibliographer applied to difficult questions concerning particular editions, issues, and states of a sig- nificant literary work. it would also be very valuable to be able to solicit input from a subject expert in determining the bounds of a concept entity (subject heading) or class entity (classification). n the experiment (my project) to explore these bewitching ideas, i have been conduct- ing an experiment. as part of my experiment, i designed a set of cataloging rules that are more frbrized than is rda in the sense that they more clearly differentiate between data applying to expression and data apply- ing to manifestation. note that there is an underlying assumption in both frbr (which defines expression quite differently from manifestation) and on my part, namely that catalogers always know whether a given piece of data applies at either the expression or the man- ifestation level. that assumption is open to questioning in the process of the experiment as well. my rules also call for creating a more hierarchical and degressive relationship between the frbr entities work, expression, manifestation, and item, such that data pertaining to the work does not need to be repeated for every expres- sion, data pertaining to the expression does not need to be repeated for every manifestation, and so forth. degressive is an old term used by bibliographers for bib- liographies that provide great detail about first editions and less detail for editions after the first. i have adapted this term to characterize my rules, according to which the cataloger begins by describing the work; any details that pertain to all expressions and manifestations of the work are not repeated in the expression and manifesta- tion descriptions. this paper would be entirely too long if i spent any more time describing the rules i am devel- oping, which can be inspected at http://myee.bol.ucla .edu. here, i would like to focus on the data-modeling process and the questions about the suitability of rdf and the semantic web for encoding our data. (by the way, i don’t seriously expect anyone to adopt my rules! they are radically different than the rules currently being applied and would represent a revolution in cata- loging practice that we may not be up to undertaking in the current economic climate. their value lies in their thought-experiment aspect and their ability to clarify what entities we can model and what entities we may not be able to model.) i am now in the process of trying to model my cataloging rules in the form of an rdf model (“rdf” as used in this paper should be considered from now on to encompass rdf schema [rdfs], web ontology language [owl], and simple knowledge organization system [skos] unless otherwise stated); this model can also be inspected at http://myee.bol .ucla.edu. in the process of doing this, i have discovered a number of areas in which i am not sure that rdf is yet sophisticated enough to deal with our data. this article is an attempt to outline some of those areas and explore whether the problems i have encountered are soluble, in other words, whether or not our data might be able to live on the semantic web eventually. i have already heard from rdf experts bruce d’arcus (miami university) and rob styles (developer of talis, as semantic web technol- ogy company), whom i cite later, but through this article i hope to reach a larger community. my research questions can be found later, but first some definitions. n definition of terms the semantic web is a way to represent knowledge; it is a knowledge-representation language that provides ways of expressing meaning that are amenable to com- putation; it is also a means of constructing knowledge- domain maps consisting of class and property axioms with a formal semantics rdf is a family of specifications for methods of modeling information that underpins the semantic web through a variety of syntax formats; an rdf metadata model is based on making statements about resources in the form of triples that consist of . the subject of the triple (e.g., “new york”); . the predicate of the triple that links the subject and the object (e.g., “has the postal abbreviation”); and . the object of the triple (e.g., “ny”). xml is commonly used to express rdf, but it is not a necessity; it can also be expressed in notation or n , for example. rdfs is an extensible knowledge-representation lan- guage that provides basic elements for the description of ontologies, also known as rdf vocabularies. using rdfs, statements are made about resources in the form of . a class (or entity) as subject of the rdf triple (e.g., “new york”); . a relationship (or semantic linkage) as predicate of the rdf triple that links the subject and the object (e.g., can bibliographic data be put directly onto the semantic web? | yee “has the postal abbreviation”); and . a property (or attribute) as object of the rdf triple (e.g., “ny”). owl is a family of knowledge representation lan- guages for authoring ontologies compatible with rdf. skos is a family of formal languages built upon rdf and designed for representation of thesauri, classification schemes, taxonomies, or subject-heading systems. n research questions actually, the full-blown semantic web may not be exactly what we need. remember that the fundamental definition of the semantic web is “a way to represent knowledge.” the semantic web is a direct descendant of the attempt to create artificial intelligence, that is, of the attempt to encode enough knowledge of the real world to allow a computer to reason about reality in a way indistinguish- able from the way a human being reasons. one of the research questions should probably be whether or not the technology developed to support the semantic web can be used to represent information rather than knowledge. fortunately, we do not need to represent all of human knowledge—we simply need to describe and index resources to facilitate their retrieval. we need to encode facts about the resources and what the resources discuss (what they are “about”), not facts about “reality.” based on our past experience, doing even this is not as simple as people think it is. the question is whether we could do what we need to do within the context of the semantic web. sometimes things that sound simple do not turn out to be so simple in the doing. my research questions are as follows: . is it possible for catalogers to tell in all cases whether a piece of data pertains to the frbr expression or the frbr manifestation? . is it possible to fit our data into rdf? given that rdf was designed to encode knowledge rather than information, perhaps it is the wrong technol- ogy to use for our purposes? . if it is possible to fit our data into rdf, is it possible to use that data to design indexes and displays that meet the objectives of the catalog (i.e., providing an efficient instrument to allow a user to find a particular work of which the author and title are known, a particular expression of a work, all of the works of an author, all of the works in a given genre or form, or all of the works on a particular subject)? as stated previously, i am not yet ready to answer these questions. i hope to find answers in the course of developing the rules and the model. in this paper, i am focusing on raising the questions about the suitability of rdf to our data that have come up in the course of my work. n other relevant projects other relevant projects include the following: . frbr, functional requirements for authority data (frad), funtional requirements for subject authority records (frsar), and frbr-object- oriented (frbroo). all are attempts to create con- ceptual models of bibliographic entities using an entity-relationship model that is very similar to the class-property model used by rdf. . various initiatives at the library of congress (lc), such as lc subject headings (lcsh) in skos, the lc name authority file in skos, the lccn permalink project to create persistent uris for bibliographic records, and initiatives to provide skos representations for vocabularies and data elements used in marc, premis, and mets. these all represent attempts to convert our exist- ing bibliographic data into uris that stand for the bibliographic entities represented by bibliographic records and authority records; the uris would then be available for experiments in putting our data directly onto the semantic web. . the dc-rda task group project to put rda data elements into rdf. as noted previously and dis- cussed further later, rda is less frbrized than my cataloging rules, but otherwise this project is very similar to mine. . dublin core’s (dc’s) work on an rdf schema. dublin core is very focused on manifestation and does not deal with expressions and works, so it is less similar to my project than is the dc-rda task groups’s project (see further discussion later). n why my project? one might legitimately ask why there is a need for a dif- ferent model than the ones already provided by frbr, frad, frsar, frbroo, rda, and dc. the frbr and rda models are still tied to the model that is implicit in our current bibliographic data in which expression and manifestation are undifferentiated. this is because publishers publish and libraries acquire and shelve mani- festations. in our current bibliographic practice, a new information technology and libraries | june bibliographic record is made for either a new manifesta- tion or a new expression. thus, in effect, there is no way for a computer to tell one from the other in our current data. despite the fact that frbr has good definitions of expression (change in content) and manifestation (mere change in carrier), it perpetuates the existing implicit model in its mapping of attributes to entities. for exam- ple, frbr maps the following to manifestation: edition statements (“ nd rev. ed.”); statements of responsibility that identify translators, editors, and illustrators; physi- cal description statements that identify illustrated edi- tions; and extent statements that differentiate expressions (the -minute version vs. the -minute version); etc. thus the frbr definition of expression recognizes that a nd revised edition is a new expression, but frbr maps the edition statement to manifestation. in my model, i have tried to differentiate more cleanly data applying to expressions from data applying to manifestations. frbr and rda tend to assume that our current bib- liographic data elements map to one and only one group entity or class. there are exceptions, such as title, which frbr and rda define at work, expression, and manifes- tation levels. however, there is a lack of recognition that, to create an accurate model of the bibliographic universe, more data elements need to be applied at the work and expression level in addition to (or even instead of) the manifestation level. in the appendix i have tried to con- trast the frbr, frad, and rda models with mine. in my model, many more data elements (properties and attri- butes) are linked to the work and expression level. after all, if the expression entity is defined as any change in work content, the work entity needs to be associated with all content elements that might change, such as the original extent of the work, the original statement of responsibil- ity, whether illustrations were originally present, whether color was originally present in a visual work, whether sound was originally present in an audiovisual work, the original aspect ratio of a moving image work, and so on. frbr also tends to assume that our current data ele- ments map to one and only one entity. in working on my model, i have come to the conclusion that this is not necessarily true. in some cases, a data element pertaining to a manifestation also pertains to the expression and the work. in other cases, the same data element is specific to that manifestation, and, in other cases, the same data ele- ment is specific to its expression. this is true of most of the elements of the bibliographic description. frad, in attempting to deal with the fact that our current cataloging rules allow a single person to have several bibliographic identities (or pseudonyms), treats person, name, and controlled access point as three separate entities or classes. i have tried to keep my model simpler and more elegant by treating only person as an entity, with preferred name and variant name as attributes or properties of that entity. frbroo is focused on the creation process for works, with special attention to the creation of unique works of art and other one-off items found in museums. thus frbroo tends to neglect the collocation of the various expressions that develop in the history of a work that is reproduced and published, such as translations, abridged editions, editions with commentary, etc. dc has concentrated exclusively on the description of manifestations and has neglected expression and work altogether. one of the tenets of semantic web development is that, once an entity is defined by a community, other communities can reuse that entity without defining it themselves. the very different definitions of the work and expression entities in the different communities described above raise some serious questions about the viability of this tenet. n assumptions it should be noted that this entire experiment is based on two assumptions about the future of human intervention for information organization. these two assumptions are based on the even bigger assumption that, even though the internet seems to be an economy based on free intel- lectual labor, and, even though human intervention for information organization is expensive (and therefore at more risk than ever), human intervention for information organization is worth the expense. n assumption : what we need is not artificial intel- ligence, but a better human–machine partnership such that humans can do all of the intellectual labor and machines can do all of the repetitive clerical labor. currently, catalogers spend too much time on the latter because of the poor design of current systems for inputting data. the univer- sal employment provided by paying humans to do the intellectual labor of building the semantic web might be just the stimulus our economy needs. n assumption : those who need structured and granular data—and the precise retrieval that results from it—to carry out research and scholarship may constitute an elite minority rather than most of the people of the world (sadly), but that talented and intelligent minority is an important one for the cul- tural and technological advancement of humanity. it is even possible that, if we did a better job of providing access to such data, we might enable the enlargement of that minority. can bibliographic data be put directly onto the semantic web? | yee n granularity and structure issues as soon as one starts to create a data model, one encoun- ters granularity or cataloger-data parsing issues. these issues have actually been with us all along as we devel- oped the data model implicit in aacr r and marc . those familiar with rda, frbr, and frad development will recognize that much of that development is directed at increasing structure and granularity in cataloger- produced data to prepare for moving it onto the semantic web. however, there are clear trade-offs in an increase in structure and granularity. more structure and more granularity make possible more powerful indexing and more sophisticated display, but more structure and more granularity are more complex and expensive to apply and less likely to be implemented in a standard fashion across all communities; that is, it is less likely that interoperable data would be produced. any switching or mapping that was employed to create interoperable data would produce the lowest common denominator (the simplest and least granular data), and once rendered interoper- able, it would not be possible for that data to swim back upstream to regain its lost granularity. data with less structure and less granularity could be easier and cheaper to apply and might have the potential to be adopted in a more standard fashion across all communities, but that data would limit the degree to which powerful indexing and sophisticated display would be possible. take the example of a personal name: currently, we demarcate surname from forename by putting the sur- name first, followed by a comma and then the forename. even that amount of granularity can sometimes pose a problem for a cataloger who does not necessarily know which part of the name is surname and which part is forename in a culture unfamiliar to the cataloger. in other words, the more granularity you desire in your data, the more often the people collecting the data are going to encounter ambiguous situations. another example: currently, we do not collect information about gender self-identification; if we were to increase the granularity of our data to gather that information, we would surely encounter situations in which the cataloger would not necessarily know if a given creator was self-defined as a female or a male or of some other gender identity. presently, if we are adding a birth and death date, whatever dates we use are all together in a $d subfield without any separate coding to indicate which date is the birth date and which is the death date (although an occa- sional “b.” or “d.” will tell us this kind of information). we could certainly provide more granularity for dates, but that would make the marc format much more complex and difficult to learn. people who dislike the marc format already argue that it is too granular and therefore requires too much of a learning curve before people can use it. for example, tennant claims that “there are only two kinds of people who believe themselves able to read a marc record without referring to a stack of manuals: a handful of our top catalogers and those on serious drugs.” how much of the granularity already in marc is used either in existing records or, even if present, is used in indexing and display software? granularity costs money, and libraries and archives are already starving for resources. granularity can only be provided by people, and people are expensive. granularity and structure also exist in tension with each other. more granularity can lead to less structure (or more complexity to retain structure along with granular- ity). in the pursuit of more granularity of data than we have now, rda, attempting to support rdf–compliant xml encoding, has been atomizing data to make it useful to computers, but this will not necessarily make the data more useful to humans. to be useful to humans, it must be possible to group and arrange (sort) the data meaning- fully, both for indexing and for display. the developers of skos refer to the “vast amounts of unstructured (i.e., human readable) information in the web,” yet labeling bits of data as to type and recording semantic relation- ships in a machine-actionable way do not necessarily provide the kind of structure necessary to make data readable by humans and therefore useful to the people the web is ultimately supposed to serve. consider the case of music instrumentation. if you have a piece of music for five guitars and one flute, and you simply code number and instrumentation without any way to link “five” with “guitars” and “one” with “flute,” you will not be able to guarantee that a person looking for music for five flutes and one guitar will not be given this piece of music in their results (see figure ). the more granular the data, the less the cataloger can build order, sequencing, and linking into the data; the coding must be carefully designed to allow the desired order, sequenc- ing, and linking for indexing and display to be possible, which might call for even more complex coding. it would be easy to lose information about order, sequencing, and linking inadvertently. actually, there are several different meanings for the term structure: . structure is an object of a record (structure of docu- ment?); for example, elings and waibel refer to “data fields . . . also referred to as elements . . . which are organized into a record by a data struc- ture.” . structure is the communications layer, as opposed to the display layer or content designation. . structure is the record, field, and subfield. . structure is the linking of bits of data together in the information technology and libraries | june form of various types of relationships. . structure is the display of data in a structured, ordered, and sequenced manner to facilitate human understand- ing. . data structure is a way of storing data in a computer so that it can be used efficiently (this is how computer programmers use the term). i hasten to add that i am definitely in favor of add- ing more structure and granularity to our data when it is necessary to carry out the fundamental objectives of our profession and of our catalogs. i argued earlier that frbr and rda are not granular enough when it comes to the distinction between data elements that apply to expression and those that apply to manifestation. if we could just agree on how to differentiate data applying to the manifestation from data applying to the expression instead of our current practice of identifying works with headings and lumping all manifestation and expression data together, we could increase the level of service we are able to provide to users a thousandfold. however, if we are not going to commit to differentiating between figure b. example of encoding of musical instrumentation at the expression level based on the above model <ycr:expinstr> <ycr:expinstrnumber> </ycr:expinstrnumber> <ycr:expinstrtype>guitars</ycr:expinstrtype> </ycr:expinstr> <ycr:expinstr> <ycr:expinstrnumber> </ycr:expinstrnumber> <ycr:expinstrtype>flute</ycr:expinstrtype> </ycr:expinstr> <rdfs:property rdf:about=“http://myee.bol.ucla.edu/ycrschema/elements/ . /expinstr”/> <rdfs:isdefinedby rdf:resource=“http://myee.bol.ucla.edu/ycrschema/elements/ . /”/> <rdfs:label xml:lang=“en”>instrumentation of musical expression</rdfs:label> <rdfs:domain rdf:resource=“http://myee.bol.ucla.edu/ycrschema#expression”/> <rdfs:range rdf:resource=“www.w .org/tr/rdf-schema#literal”/> <rdfs:subpropertyof rdf:resource=“http://myee.bol.ucla.edu/ycrschema#expdesc”/> <rdfs:property rdf:about=“http://myee.bol.ucla.edu/ycrschema/elements/ . /expinstrnumber”/> <rdfs:isdefinedby rdf:resource=“http://myee.bol.ucla.edu/ycrschema/elements/ . /”/> <rdfs:label xml:lang=“en”>original instrumentation of musical expression—number of a particular instrument</ rdfs:label> <rdfs:domain rdf:resource=“http://myee.bol.ucla.edu/ycrschema#expression”/> <rdfs:range rdf:resource=“www.w .org/tr/rdf-schema#literal”/> <rdfs:subpropertyof rdf:resource=“http://myee.bol.ucla.edu/ycrschema#expinstr”/> <rdfs:property rdf:about=“http://myee.bol.ucla.edu/ycrschema/elements/ . /expinstrtype”/> <rdfs:isdefinedby rdf:resource=“http://myee.bol.ucla.edu/ycrschema/elements/ . /”/> <rdfs:label xml:lang=“en”>original instrumentation of musical expression—type of instrument</rdfs:label> <rdfs:domain rdf:resource=“http://myee.bol.ucla.edu/ycrschema#expression”/> <rdfs:range rdf:resource=“www.w .org/tr/rdf-schema#literal”/> <rdfs:subpropertyof rdf:resource=“http://myee.bol.ucla.edu/ycrschema#expinstr”/> figure a. extract from yee rdf model that illustrates one technique for modeling musical instrumentation at the expression level (using a blank node to group repeated number and instrument type) can bibliographic data be put directly onto the semantic web? | yee expression and manifestation, it would be more intellec- tually honest for frbr and rda to take the less granular path of mapping all existing bibliographic data to mani- festation and expression undifferentiated, that is, to use our current data model unchanged and state this openly. i am not in favor of adding granularity for granularity’s sake or for the sake of vague conceptions of possible future use. granularity is expensive and should be used only in support of clear and fundamental objectives. n the goal: efficient displays and indexes my main concern is that we model and then structure the data in a way that allows us to build the complex displays that are necessary to make catalogs appear simple to use. i am aware that the current orthodoxy is that recording data should be kept completely separate from indexing and display (“the applications layer”). because i have spent my career in a field in which catalog records are indexed and displayed badly by systems people who don’t seem to understand the data contained in them, i am a skeptic. it is definitely possible to model and struc- ture data in such a way that desired displays and indexes are impossible to construct. i have seen it happen! the lc working group report states that “it will be recognized that human users and their needs for display and discovery do not represent the only use of bibliographic metadata; instead, to an increasing degree, machine applications are their primary users.” my fear is that the underlying assumption here is that users need to (and can) retrieve the single perfect record. this will never be true for bibliographic metadata. users will always need to assemble all relevant records (of all kinds) as precisely as possible and then browse through them before making a decision about which resources to obtain. this is as true in the semantic web—where “records” can be conceived of as entity or class uris—as it is in the world of marc–encoded metadata. some of the problems that have arisen in the past in trying to index bibliographic metadata for humans are connected to the fact that existing systems do not group all of the data related to a particular entity effectively, such that a user can use any variant name or any combi- nation of variant names for an entity and do a successful search. currently, you can only look for a match among two or more keywords within the bounds of a single manifestation-based bibliographic record or within the bounds of a single heading, minus any variant terms for that entity. thus, when you do a keyword search for two keywords, for example, “clemens” and “adventures,” you will retrieve only those manifestations of mark twain’s adventures of tom sawyer that have his real name (clemens) and the title word “adventures” co-occurring within the bounded space created by a single manifes- tation-based bibliographic record. instead, the preferred forms and the variant forms for a given entity need to be bounded for indexing such that the keywords the user employs to search for that entity can be matched using co-occurrence rules that look for matches within a single bounded space representing the entity desired. we will return to this problem in the discussion of issue in the later section “rdf problems encountered.” the most complex indexing problem has always proven to be the grouping or bounding of data related to a work, since it requires pulling in all variants for the creator(s) of that work as well. otherwise, a user who searches for a work using a variant of the author’s name and a variant of the title will continue to fail (as they do in all current opacs), even when the desired work exists in the catalog. if we could create a uri for the adventures of tom sawyer that included all variant names for the author and all variant titles for the work (including the variant title tom sawyer), the same keyword search described above (“clemens” and “adventures”) could be made to retrieve all manifestations and expressions of the adventures of tom sawyer, instead of the few isolated manifestations that it would retrieve in current catalogs. we need to make sure that we design and structure the data such that the following displays are possible: n display all works by this author in alphabeti- cal order by title with the sorting element (title) appearing at the top of each work displayed. n display all works on this subject in alphabetical order by principal author and title (with principal author and title appearing at top of each work dis- played), or title if there is no principal author (with title appearing at top of each work displayed). we must ensure that we design and structure the data in such a way that our structure allows us to create subgroups of related data, such as instrumentation for a piece of music (consisting of a number associated with each particular instrument), place and related publisher for a certain span of dates on a serial title change record, and the like. n which standards will carry out which functions? currently, we have a number of different standards to carry out a number of different functions; we can speculate about how those functions might be allocated in a new semantic web–based dispensation, as shown in table . in table , data structure is taken to mean what a record represents or stands for; traditionally, a record has represented an expression (in the days of hand- information technology and libraries | june press books) or a manifestation (ever since reproduction mechanisms have become more sophisticated, allowing an explosion of reproductions of the same content in different formats and coming from different distribu- tors). rda is record-neutral; rdf would allow uris to be established for any and all of the frbr levels; that is, there would be a uri for a particular work, a uri for a particular expression, a uri for a particular manifesta- tion, and a uri for a particular item. note that i am not using data structure in the sense that a computer pro- grammer does (as a way of storing data in a computer so that it can be used efficiently). currently, the encoding of facts about entity relation- ships (see table ) is carried out by matching data-value character strings (headings or linking fields using issns and the like) that are defined by the lc/naco author- ity file (following aacr r rules), lcsh (following rules in the subject cataloging manual), etc. in the future, this function might be carried out by using rdf to link the uri for a resource to the uri for a data value. display rules (see table ) are currently defined by isbd and aacr r but widely ignored by systems, which frequently truncate bibliographic records arbitrarily in displays, supply labels, and the like; rda abdicates responsibility, pushing display out of the cataloging rules. the general principle on the web is to divorce data from display and allow anyone to display the data any way they want. display is the heart of the objects (or goals) of cataloging: the point is to display to the user the works of an author, the editions of a work, or the works on a subject. all of these goals only can be met if complex, high-quality displays can be built from the data created according to the data model. indexing rules (see table ) were once under the control of catalogers (in book and card catalogs) in that users had to navigate through headings and cross-references to find table . possible reallocation of current functions in a new semantic web–based dispensation function current future? data content, or content guidelines (rules for providing data in a particular element) defined by aacr r and marc defined by rda and rdf/rdfs/ owl/skos data elements defined by isbd–based aacr r and marc defined by rda and rdf/rdfs/ owl/skos data values defined by lc/naco authority file, lcsh, marc coded data values, etc. defined as ontologies using rdf/ rdfs/owl/skos encoding or labeling of data elements for machine manipulation; same as data format? defined by iso –based marc defined by rdf/rdfs/xml data structure (i.e., what a record stands for) defined by aacr r and marc ; also frbr? defined by rdf/rdfs/owl/ skos schematization (constraint on structure and content) marc , mods, dcmi abstract model defined by rdf/rdfs/owl/ skos encoding of facts about entity relationships carried out by matching data value strings (headings found in lc/naco authority file and lcsh, issn’s, and the like) carried out by rdf/rdfs/owl/ skos in the form of uri links display rules ils software, formerly isbd– based aacr r (“application layer”) or yee rules indexing rules ils software sparql, “application layer,” or yee rules can bibliographic data be put directly onto the semantic web? | yee what they wanted; currently indexing is in the hands of system designers who prefer to provide keyword index- ing of bibliographic (i.e., manifestation-based) records rather than provide users with access to the entities they are really interested in (works, authors and subjects), all represented currently by authority records for head- ings and cross-references. rda abdicates responsibility, pushing indexing concerns completely out of the catalog- ing rules. the general principle on the web is to allow resources to be indexed by any web search engines that wish to index them. current web data is not structured at all for either indexing or display. i would argue that our interest in the semantic web should be focused on whether or not it will support more data structure—as well as more logic in that data structure—to support better indexes and better displays than we have now in manifestation-based ils opacs. crucial to better indexing than we have ever had before are the co-occurrence rules for keyword indexing, that is, the rules for when a co-occurrence of two or more keywords should produce a match. we need to be able to do a keyword search across all possible variant names for the entity of interest, and the entity of interest for the average catalog user is much more likely to be a particular work than to be a particular manifestation. unfortunately, catalog-use studies only have studied so-called known-item searches without investigating whether a known-item searcher was looking for a par- ticular edition or manifestation of a work or was simply looking for a particular work in order to make a choice as to edition or manifestation once the work was found. however, common sense tells us that it is a rare user who approaches the catalog with prior knowledge about all published editions of a given work. the more com- mon situation is surely one in which a user desires to read a particular shakespeare play or view a particular david lean film and discovers that the desired work exists in more than one expression or manifestation only after searching the catalog. we need to have the keyword(s) in our search for a particular work co-occur within a bounded space that encompasses all possible keywords that might refer to that particular work entity, including both creator and title keywords. notice in table the unifying effect that rdf could potentially have; it could free us from the use of multiple standards that can easily contradict each other, or at least not live peacefully together. examples are not hard to find in the current environment. one that has cropped up in the course of rda development concerns family names. presently the rules for naming families are dif- ferent depending on whether the family is the subject of a work (and established according to lcsh) or whether the family is responsible for a collection of papers (and established according to rda). n types of data rda has blurred the distinctions among certain types of data, apparently because there is a perception that on the semantic web the same piece of data needs to be coded only once, and all indexing and display needs can be supported from that one piece of data. i question that assumption on the basis of my experience with biblio- graphic cataloging. all of the following ways of encod- ing the same piece of data can still have value in certain circumstances: n transcribed; in rdf terms, a literal (i.e., any data that is not a uri, a constant value). transcribed data is data copied from an item being cataloged. it is valuable for providing access to the form of the name used on a title page and is particularly useful for people who use pseudonyms, corporate bodies that change name, and so on. transcribed data is an important part of the historical record and not just for off-line materials; it can be a historical record of changing data on notoriously fluid webpages. n composed; in rdf terms, also a literal. composed data is information composed by a cataloger on the basis of observation of the item in hand; it can be valuable for historical purposes to know which data was composed. n supplied; in rdf terms, also a literal. supplied data is information supplied by a cataloger from outside sources; it can be valuable for historical purposes to know which data was supplied and from which outside sources it came. n coded; in rdf, represented by a uri. coded data would likely transform on the semantic web into links to ontologies that could provide normalized, human-readable identification strings on demand, thus causing coded and normalized data to merge into one type of data. is it not possible, though, that the coded form of normalized data might continue to provide for more efficient searching for computers as opposed to humans? coded data also has great cross-cultural value, since it is not as language-dependent as literals or normalized headings. n normalized headings (controlled headings); in rdf, represented by a uri. normalized or controlled headings are still necessary to provide users with coherent, ordered displays of thousands of entities that all match the user’s search for a particular entity (work, author, subject, etc.). the reason google displays are so hideous is that, so far, the data searched lacks any normalized display data. if variant language forms of the name for an entity information technology and libraries | june are linked to an entity uri, it should be possible to supply headings in the language and script desired by a particular user. n the rdf model those who have become familiar with frbr over the years will probably not find it too difficult to transition from the frbr conceptual model to the rdf model. what frbr calls an “entity,” rdf calls a “subject” and rdfs calls a “class.” what frbr calls an “attribute,” rdf calls an “object” and rdfs calls a “property.” what frbr calls a “relationship,” rdf calls a “predicate” and rdfs calls a “relationship” or a “semantic linkage” (see table ). the difficulty in any data-modeling exercise lies in deciding what to treat as an entity or class and what to treat as an attribute or property. the authors of frbr decided to create a class called expression to deal with any change in the content of a work. when frbr is applied to serials, which change content with every issue, the model does not work well. in my model, i found it useful to create a new entity at the manifestation level, the serial title, to deal with the type of change that is more relevant to serials, the change in title. i also created another new entity at the manifestation level, title-manifestation, to deal with a change of title in a nonserial work that is not asso- ciated with a change in content. one hundred years ago, this entity would have been called title-edition. i am also in the process of developing an entity at the expression level—surrogate—to deal with reproductions of original artworks that need to inherit the qualities of the original artwork they reproduce without being treated as an edi- tion of that original artwork, which ipso facto is unique. these are just examples of cases in which it is not that easy to decide on the classes or entities that are necessary to accurately model bibliographic information. see the appendix for a complete comparison of the classes and entities defined in four different models: frbr, frad, rda, and the yee cataloging rules (ycr). the appendix also shows variation among these models concerning whether a given data element is treated as a class/entity or as an attribute/property. the most notable examples are name and preferred access point, which are treated as classes/entities in frad, as attributes in frbr and ycr, and as both in rda. n rdf problems encountered my goal for this paper is to institute discussion with data modelers about which problems i observed are insoluble and which are soluble: . is there an assumption on the part of semantic web developers that a given data element, such as a publisher name, should be expressed as either a literal or using a uri (i.e., con- trolled), but never both? cataloging is rooted in humanistic practices that require careful recording of evidence. there will always be value in distinguishing and labeling the following types of data: n copied as is from an artifact (transcribed) n supplied by a cataloger n categorized by a cataloger (controlled) tim berners-lee (the father of the internet and the semantic web) emphasizes the importance of record- ing not just data but also its provenance for the sake of authenticity. for many data elements, therefore, it will be important to be able to record both a literal (tran- scribed or composed form or both) and a uri (controlled form). is this a problem in rdf? as a corollary, if any data that can be given a uri cannot also be represented by a literal (transcribed and composed data, or one or the other), it may not be possible to design coherent, readable displays of the data describing a particular entity. among other things, cataloging is a discursive writing skill. does rdf require that all data be represented only once, either by a literal or by a uri? or is it perhaps possible that data that has a uri could also have a transcribed or composed form as a property? perhaps it will even be possible to store multiple snapshots of online works that change over time to document variant forms of a name for works, persons, and so on. . will the internet ever be fast enough to assemble the equivalent of our current records from a collection of hundreds or even thousands of uris? in rdf, links are one-to-one rather than one-to-many. this leads to a great prolifera- tion of reciprocal links. the more granularity there is in the data, the more linking is necessary to ensure that atomized data elements are linked together. potentially, every piece of data describing a particular entity could be represented by a uri leading out to a skos list of data values. the number of links necessary to pull together table . the frbr conceptual model translated into rdf and rdfs frbr rdf rdfs entity subject class attribute object property relationship predicate relationship/ semantic linkage can bibliographic data be put directly onto the semantic web? | yee all of the data just to describe one manifestation could become astronomical, as could the number of one-to-one links necessary to create the appearance of a one-to-many link, such as the link between an author and all the works of an author. is the internet really fast enough to assemble a record from hundreds of uris in a reasonable amount of time? given the often slow network throughput typical of many of our current internet connections, is it really practical to expect all of these pieces to be pulled together efficiently to create a single display for a single user? we yet may feel nostalgia for the single manifestation-based record that already has all of the relevant data in it (no assembly required). bruce d’arcus points out, however, that i think if you’re dealing with rdf, you wouldn’t neces- sarily be gathering these data in real-time. the uris that are the targets for those links are really just global identifiers. how you get the triples is a separate matter. so, for example, in my own personal case, i’m going to put together an rdf store that is populated with data from a variety of sources, but that data popula- tion will happen by script, and i’ll still be querying a single endpoint, where the rdf is stored in a relational database. in other words, d’arcus essentially will put them all in one place, or in one database that “looks” from a uri perspective to be “one place” where they’re already gathered. . is rdf capable of dealing with works that are identified using their creators? we need to treat author as both an entity in its own right and as a property of a work, and in many cases the latter is the more important function for user service. lexical labels, or human-readable identi- fiers for works that are identified using both the principal author and the title, are particularly problematic in rdf given that the principal author is an entity in its own right. is rdf capable of supporting the indexing neces- sary to allow a user to search using any variant of the author’s name and any variant of the title of a work in combination and still retrieve all expressions and mani- festations of that work, given that author will have a uri of its own, linked by means of a relationship link to the work uri? is rdf capable of supporting the display of a list of one thousand works, each identified by principal author, in order first by principal author, then by title, then by publication date, given that the preferred heading for each principal author would have to be assembled from the uri for that principal author and the preferred title for each work would have to be assembled from the uri for that work? for fear that this will not, in fact, be pos- sible, i have put a human-readable work-identifier data element into my model that consists of principal author and title when appropriate, even though that means the preferred name of the principal author may not be able to be controlled by the entity record for the principal author. any guidance from experienced data modelers in this regard would be appreciated. according to bruce d’arcus, this is purely an inter- face or application question that does not require a solu- tion at the data layer. since we have never had interfaces or applications that would do this correctly, even though the data is readily available in authority records, i am skeptical about this answer! perhaps bruce’s suggestion under item of designat- ing a sortname property for each entity is the solution here as well. my human-readable work identifier con- sisting of the name of the principal creator and uniform title of work could be designated the sortname poperty for the work. it would have to be changed whenever the preferred form of the name for the principal creator changed, however. . do all possible inverse relationships need to be expressed explicitly, or can they be inferred? my model is already quite large, and i have not yet defined the inverse of every property as i really should to have a correct rdf model. in other words, for every property there needs to be an inverse property; for example, the property iscreatorof needs to have the inverse property iscreatedby; thus “twain” has the property iscreatorof, while “adventures of tom sawyer” has the property iscreatedby. perhaps users and inputters will not actually have to see the huge, complex rdf data model that would result from creating all the inverse relationships, but those who maintain the model will have to deal with a great deal of complexity. however, since i’m not a programmer, i don’t know how the complexity of rdf compares to the complexity of existing ils software. . can rdf solve the problems we are having now because of the lack of transitivity or inheritance in the data models that underlie current ilses, or will rdf merely perpetuate these problems? we have problems now with the data models that underlie our current ilses because of the inability of these models to deal with hierarchical inheritance, such that whatever is true of an entity in the hierarchy is also true of every entity below that entity in the hierarchy. one example is that of cross-references to a parent corporate body that should be held to apply to all subdivisions of that corporate body but never are in existing ils systems. there is a cross-reference from “fbi” to “united states. federal bureau of investigation,” but not from “fbi counterterrorism division” to “united states. federal bureau of investigation. counterterrorism division.” for that reason, a search in any opac name index for “fbi counterterrorism division” will fail. we need systems that recognize that data about a parent corporate body is relevant to all subdivisions of that parent body. we need systems that recognize that data about a work is relevant to all expressions and manifestations of that work. rdf allows you to link a work to an expression information technology and libraries | june and an expression to a manifestation, but i don’t believe it allows you to encode the information that everything that is true of the work is true of all of its expressions and manifestations. rob styles seems to confirm this: “rdf doesn’t have hierarchy. in computer science terms, it’s a graph, not a tree, which means you can connect anything to anything else in any direction.” of course, not all links should be this kind of tran- sitive or inheritance link. one expression of work a is linked to another expression of work a by links to work a, but whatever is true of one of those expressions is not necessarily true of the other; one may be illustrated, for example, while the other is not. whatever is true of one work is not necessarily true of another work related to it by related work link. it should be recognized that bibliographic data is rife with hierarchy. it is one of our major tools for expressing meaning to our users. corporate bodies have corporate subdivisions, and many things that are true for the par- ent body also are true for its subdivisions. subjects are expressed using main headings and subject subdivisions, and many things that are true for the main heading (such as variant names) also are true for the heading combined with one of its subdivisions. geographic areas are con- tained within larger geographic areas, and many things that are true of the larger geographic area also are true for smaller regions, counties, cities, etc., contained within that larger geographic area. for all these reasons, i believe that, to do effective displays and indexes for our biblio- graphic data, it is critical that we be able to distinguish between a hierarchical relationship and a nonhierarchical relationship. . to recognize the fact that the subject of a book or a film could be a work, a person, a concept, an object, an event, or a place (all classes in the model), is there any reason we cannot define subject itself as a property (a relationship) rather than a class in its own right? in my model, all subject properties are defined as having a domain of resource, meaning there is no constraint as to the class to which these subject properties apply. i’m not sure if there will be any fall-out from that modeling decision. . how do we distinguish between the corporate behavior of a jurisdiction and the subject behavior of a geographical loca- tion? sometimes a place is a jurisdiction and behaves like a corporate body (e.g., united states is the name of the government of the united states). sometimes place is a physical location in which something is located (e.g., the birds discussed in a book about the birds of the united states). to distinguish between the corporate behavior of a jurisdiction and the subject behavior of a geographical location, i have defined two different classes for place: place as jurisdictional corporate body and place as geographic area. will this cause problems in the model? will there be times when it prevents us from making elegant general- izations in the model about place per se? there is a similar problem with events. some events are corporate bodies (e.g., conferences that publish papers) and some are a kind of subject (e.g., an earthquake). i have defined two different classes for event: conference or other event as corporate body creator and event as subject. . what is the best way to model a bound-with or an issued- with relationship, or a part–whole relationship in which the whole must be located to obtain the part? the bound-with relationship is actually between two items containing two different works, while the issued-with relationship is between two manifestations containing two different works (see figure ). is this a work-to-work relation- ship? will designating it a work-to-work relationship cause problems for indicating which specific items or manifestation-items of each work are physically located in the same place? this question may also apply to those part–whole relationships in which the part is physically contained within the whole and both are located in the same place (sometimes known as analytics). one thing to bear in mind is that in all of these cases the relationship between two works does not hold between all instances of each work; it only holds for those particular instances that are contained in the particular manifestation or item that is bound with, issued with, or part of the whole. however, if the relationship is modeled as a work- - manifestation to work- -manifestation relationship, or a work- -item to work- -item relationship,, care must be taken in the design of displays to pull in enough infor- mation about the two or more works so as not to confuse the user. . how do we express the arrangement of elements that have a definite order? i am having trouble imagining how to encode the ordering of data elements that make up a larger element, such as the pieces of a personal name. this is really a desire to control the display of those atom- ized elements so that they make sense to human beings rather than just to machines. could one define a property such as natural language order of forename, surname, middle name, patronymic, matronymic and/or clan name of a person given that the ideal order of these elements might vary from one person to another? could one define proper- ties such as sorting element , sorting element , sorting element , etc., and assign them to the various pieces that will be assembled to make a particular heading for an entity, such as an lcsh heading for a historical period? (depending on the answer to the question in item , it may or may not be possible to assign a property to a property in this fashion.) are there standard sorting rules we need to be aware of (in unicode, for example)? are there other rdf techniques available to deal with sorting and arrangement? bruce d’arcus suggests that, instead of coding the name parts, it would be more useful to designate sort- name properties; might it not be necessary to designate a sortname property for each variant name, as well, can bibliographic data be put directly onto the semantic web? | yee for cases in which variants need to appear in sorted displays? and wouldn’t these sortname properties com- plicate maintenance over time as preferred and variant names changed? . how do we link related data elements in such a way that effective indexing and displays are possible? some examples: number and kind of instrument (e.g., music written for two oboes and three guitars); multiple publishers, fre- quencies, subtitles, editors, etc., with date spans for a serial title change (or will it be necessary to create a new manifestation for every single change in subtitle, pub- lisher name, place of publication, etc?). the assumption seems to be that there will be no repeatable data ele- ments. based on my somewhat limited experience with rdf, it appears that there are record equivalents (every data element—property or relationship—pertaining to a particular entity with a uri), but there are no field or subfield equivalents that allow the sublinking of related pieces of data about an entity. indeed, rob styles goes so far as to argue that ultimately there is no notion of a “record” in rdf. it is possible that blank nodes might be able to fill in for fields and subfields in some cases for grouping data, but there are dangers involved in their use. to a cataloger, it looks as though the plan is for rdf data to float around loose without any requirement that there be a method for pulling it together into coherent displays designed for human beings. . can a property have a property in rdf? as an exam- ple of where it might be useful to define a property of a property, robert maxwell suggests that date of publication is really an attribute (property) of the published by rela- tionship (another property). another example: in my model, a variant title for a serial is a property. can that property itself have the property type of variant title to encompass things like spine title, key title, etc.? another example appeared in item , in which it is suggested that it might be desirable to assign sort-element properties to the various elements of a name property. . how do we document record display decisions? there is no way to record display decisions in rdf itself; it is completely display-neutral. we could not safely commit to a particular rdf–based data model until a significant amount of sample bibliographic data had been created and open-source indexing and display software had been designed and user-tested on that data. it may be that we will need to supplement rdf with some other encoding mechanism that allows us to record display decisions along with the data. current cataloging rules are about display as much as they are about content designation. isbd concerns the order in which the elements should be displayed to humans. the cataloging objectives con- cern display to users of such entity groups as the works of an author, the editions of a work, and the works on a subject. . can all bibliographic data be reduced to either a class or a property with a finite list of values? another way to put this is to ask if all that catalogers do could be reduced to a set of pull-down menus. cataloging is the art of writing discursive prose as much as it is the ability to select the correct value for a particular data element. we must deal with ambiguous data (presented by joe blow could mean that joe created the entire work, produced it, distributed it, sponsored it, or merely funded it). we must sometimes record information without knowing its exact meaning. we must deal with situations that have not been antici- pated in advance. it is not possible to list every possible kind of data and every possible value for each type of figure . examples of part–whole relationships. how might these be best expressed in rdf? issued-with relationship a copy of charlie chaplin’s film the immigrant can be found on a videodisc compilation called charlie chaplin, the early years along with two other chaplin films. this compilation was published and collected by many different libraries and media centers. if a user wants to view this copy of the immigrant, he or she will first have to locate charlie chaplin, the early years, then look for the desired film at the beginning of the first videodisc in the set. the issued-with rela- tionship between the immigrant and the other two films on charlie chaplin, the early years is currently expressed in the bibliographic record by means of a “with” note: first on charlie chaplin, the early years, v. ( min.) with: the count – easy street. bound-with relationship the university of california, los angeles film & television archive has acquired a reel of mm. film from a collector who strung five warner bros. car- toons together on a single reel of film. we can assume that no other archive, library, or media collection will have this particular compilation of cartoons, so the relationship between the five cartoons is purely local in nature. however, any user at the film & television archive who wishes to view one of these cartoons will have to request a viewing appointment for the entire reel and then find the desired cartoon among the other four on the reel. the bound-with relation- ship among these cartoons is currently expressed in a holdings record by means of a “with” note: fourth on reel with: daffy doodles – tweety pie – i love to singa – along flirtation walk. information technology and libraries | june data up front before any data is gathered. it will always be necessary to provide a plain-text escape hatch. the bibliographic world is a complex, constantly changing world filled with ambiguity. n what are the next steps? in a sense, this paper is a first crude attempt at locating unmapped territory that has not yet been explored. if we were to decide as a community that it would be valu- able to move our shared cataloging activities onto the semantic web, we would have a lot of work ahead of us. if some of the rdf problems described above are insolu- ble, we may need to work with semantic web developers to create a more sophisticated version of rdf that can handle the transitivity and complex linking required by our data. we will also need to encourage a very complex existing community to evolve institutional structures that would enable a more efficient use of the internet for the sharing of cataloging and other metadata creation. this is not just a technological problem, but also a political one. in the meantime, the experiment continues. let the think- ing and learning begin! references and notes . “notation , or n as it is more commonly known, is a shorthand non–xml serialization of resource description framework models, designed with human-readability in mind: n is much more compact and readable than xml rdf nota- tion. the format is being developed by tim berners-lee and oth- ers from the semantic web community.” wikipedia, “notation ,” http://en.wikipedia.org/wiki/notation_ (accessed feb. , ). . frbr review group, www.ifla.org/vii/s /wgfrbr/; frbr review group, franar (working group on functional requirements and numbering of authority records), www .ifla.org/vii/d /wg-franar.htm; frbr review group, frsar (working group, functional requirements for subject authority records), www.ifla.org/vii/s /wgfrsar.htm; frbroo, frbr review group, working group on frbr/crm dialogue, www .ifla.org/vii/s /wgfrbr/frbr-crmdialogue_wg.htm. . library of congress, response to on the record: report of the library of congress working group on the future of bib- liographic control (washington, d.c.: library of congress, ): , , , www.loc.gov/bibliographic-future/news/lcwgrpt response_dm_ .pdf (accessed mar. , ). . ibid., . . ibid., . . dublin core metadata initiative, dcmi/rda task group wiki, http://www.dublincore.org/dcmirdataskgroup/ (accessed mar. , ). . mikael nilsson, andy powell, pete johnston, and ambjorn naeve, expressing dublin core metadata using the resource description framework (rdf), http://dublincore.org/ documents/ / / /dc-rdf/ (accessed mar. , ). . see for example table . in frbr, which maps to mani- festation every kind of data that pertains to expression change with the exception of language change. ifla study group on the functional requirements for bibliographic records, func- tional requirements for bibliographic records (munich: k. g. saur, ): , http://www.ifla.org/vii/s /frbr/frbr.pdf (accessed mar. , ). . roy tennant, “marc must die,” library journal , no. (oct. , ): . . w c, skos simple knowledge organization system refer- ence, w c working draft august , http://www.w .org/ tr/skos-reference/ (accessed mar. , ). . the extract in figure is taken from my complete rdf model, which can be found at http://myee.bol.ucla.edu/ ycrschemardf.txt. . mary w. elings and gunter waibel, “metadata for all: descriptive standards and metadata sharing across libraries, archives and museums,” first monday , no. (mar. , ), http://www.uic.edu/htbin/cgiwrap/bin/ojs/index.php/fm/ article/view/ / (accessed mar. , ). . oclc, a holdings primer: principles and standards for local holdings records, nd ed. (dublin, ohio: oclc, ), , http:// www.oclc.org/us/en/support/documentation/localholdings/ primer/holdings% primer% .pdf (accessed mar. , ). . the library of congress working group, on the record: report of the library of congress working group on the future of bibliographic control (washington, d.c.: library of congress, ): , http:// www.loc.gov/bibliographic-future/news/lcwg-ontherecord -jan -final.pdf (accessed mar. , ). . talis, sir tim berners-lee talks with talis about the seman- tic web: transcript of an interview recorded on february , http://talis-podcasts.s .amazonaws.com/twt _timbl .html (accessed mar. , ). . bruce d’arcus, e-mail to author, mar. , . . ibid. . rob styles, e-mail to author, mar. , . . bruce d’arcus, e-mail to author, mar. , . . rob styles, e-mail to author, mar. , . . w c, “section . , structured property values and blank nodes,” in rdf primer: w c recommendation february , http://www.w .org/tr/rdf-primer/#structuredproperties (accessed mar. , ). . robert maxwell, frbr: a guide for the perplexed (chicago: ala, ). can bibliographic data be put directly onto the semantic web? | yee entities/classes in rda, frbr, frad compared to yee cataloging rules (ycr) rda, frbr, and frad ycr group : work work group : expression expression surrogate group : manifestation manifestation title-manifestation serial title group : item item group : person person fictitious character performing animal group : corporate body corporate body corporate subdivision place as jurisdictional corporate body conference or other event as corporate body creator jurisdictional corporate subdivision family (rda and frad only) group : concept concept group : object object group : event event or historical period as subject group : place place as geographic area discipline genre/form name identifier controlled access point rules (frad only) agency (frad only) appendix. entity/class and attribute/property comparisons information technology and libraries | june attributes/properties in frbr compared to frad model entity frbr frad work title of the work form of work date of the work other distinguishing characteristics intended termination intended audience context for the work medium of performance (musical work) numeric designation (musical work) key (musical work) coordinates (cartographic work) equinox (cartographic work) form of work date of the work medium of performance subject of the work numeric designation key place of origin of the work original language of the work history other distinguishing characteristic expression title of the expression form of expression date of expression language of expression other distinguishing characteristics extensibility of expression revisability of expression extent of the expression summarization of content context for the expression critical response to the expression use restrictions on the expression sequencing pattern (serial) expected regularity of issue (serial) expected frequency of issue (serial) type of score (musical notation) medium of performance (musical notation or recorded sound) scale (cartographic image/object) projection (cartographic image/object) presentation technique (cartographic image/object) representation of relief (cartographic image/object) geodetic, grid, and vertical measurement (cartographic image/ object) recording technique (remote sensing image) special characteristic (remote sensing image) technique (graphic or projected image) form of expression date of expression language of expression technique other distinguishing characteristic surrogate can bibliographic data be put directly onto the semantic web? | yee model entity frbr frad manifestation title of the manifestation statement of responsibility edition/issue designation place of publication/distribution publisher/distributor date of publication/distribution fabricator/manufacturer series statement form of carrier extent of the carrier physical medium capture mode dimensions of the carrier manifestation identifier source for acquisition/access authorization terms of availability access restrictions on the manifestation typeface (printed book) type size (printed book) foliation (hand-printed book) collation (hand-printed book) publication status (serial) numbering (serial) playing speed (sound recording) groove width (sound recording) kind of cutting (sound recording) tape configuration (sound recording) kind of sound (sound recording) special reproduction characteristic (sound recording) colour (image) reduction ratio (microform) polarity (microform or visual projection) generation (microform or visual projection) presentation format (visual projection) system requirements (electronic resource) file characteristics (electronic resource) mode of access (remote access electronic resource) access address (remote access electronic resource) edition/issue designation place of publication/distribution publisher/distributor date of publication/distribution form of carrier numbering title-manifestation serial title item item identifier fingerprint provenance of the item marks/inscriptions exhibition history condition of the item treatment history scheduled treatment access restrictions on the item location of item attributes/properties in frbr compared to frad (cont.) information technology and libraries | june model entity frbr frad person name of person dates of person title of person other designation associated with the person dates associated with the person title of person other designation associated with the person gender place of birth place of death country place of residence affiliation address language of person field of activity profession/occupation biography/history fictitious character performing animal corporate body name of the corporate body number associated with the corporate body place associated with the corporate body date associated with the corporate body other designation associated with the corporate body place associated with the corporate body date associated with the corporate body other designation associated with the corporate body type of corporate body language of the corporate body address field of activity history corporate subdivision place as jurisdictional corporate body conference or other event as corporate body creator jurisdictional corporate subdivision family type of family dates of family places associated with family history of family concept term for the concept type of concept object term for the object type of object date of production place of production producer/fabricator physical medium event term for the event date associated with the event place associated with the event attributes/properties in frbr compared to frad (cont.) can bibliographic data be put directly onto the semantic web? | yee model entity frbr frad place term for the place coordinates other geographical information discipline genre/form name type of name scope of usage dates of usage language of name script of name transliteration scheme of name identifier type of identifier identifier string suffix controlled access point type of controlled access point status of controlled access point designated usage of controlled access point undifferentiated access point language of base access point script of base access point script of cataloguing transliteration scheme of base access point transliteration scheme of cataloguing source of controlled access point base access point addition rules citation for rules rules identifier agency name of agency agency identifier location of agency attributes/properties in frbr compared to frad (cont.) information technology and libraries | june attributes/properties in rda compared to ycr model entity rda ycr work title of the work form of work date of work place of origin of work medium of performance numeric designation key signatory to a treaty, etc. other distinguishing characteristic of the work original language of the work history of the work identifier for the work nature of the content coverage of the content coordinates of cartographic content equinox epoch intended audience system of organization dissertation or theses information key identifier for work language-based identifier (preferred lexical label) variant language-based identifier (alternate lexical label) language-based identifier (preferred lexical label) for work language-based identifier for work (preferred lexical label) identified by principalcreator in combination with uniform title language-based identifier (preferred lexical label) for work identified by title alone (uniform title) supplied title for work variant title for work original language of work responsibility for work original publication statement of work dates associated with work original publication/release/broadcast date of work copyright date of work creation date of work date of first recording of a work date of first performance of a work finding date of naturally occurring object original publisher/distributor/broadcaster of work places associated with work original place of publication/distribution/broadcasting for work country of origin of work place of creation of work place of first recording of work place of first performance of work finding place of naturally occurring object original method of publication/distribution/broadcast of work serial or integrating work original numeric and/or alphabetic designations—beginning serial or integrating work original chronological designations— beginning serial or integrating work original numeric and/or alphabetic designations—ending serial or integrating work original chronological designations— ending encoding of content of work genre/form of content of work original instrumentation of musical work instrumentation of musical work—number of a particular instrument instrumentation of musical work—type of instrument original voice(s) of musical work voice(s) of musical work—number of a particular type of voice voice(s) of musical work—type of voice original key of musical work numeric designation of musical work coordinates of cartographic work equinox of cartographic work original physical characteristics of work original extent of work original dimensions of work mode of issuance of work can bibliographic data be put directly onto the semantic web? | yee model entity rda ycr work (cont.) original aspect ratio of moving image work original image format of moving image work original base of work original materials applied to base of work work summary work contents list custodial history of work creation of archival collection censorship history of work note about relationship(s) to other works expression content type date of expression language of expression other distinguishing characteristic of the expression identifier for the expression summarization of the content place and date of capture language of the content form of notation accessibility content illustrative content supplementary content colour content sound content aspect ratio format of notated music medium of performance of musical content duration performer, narrator, and/or presenter artistic and/or technical credits scale projection of cartographic content other details of cartographic content awards key identifier for expression language-based identifier (preferred lexical label) for expression variant title for expression nature of modification of expression expression title expression statement of responsibility edition statement scale of cartographic expression projection of cartographic expression publication statement of expression place of publication/distribution/release/broadcasting for expression place of recording for expression publisher/distributor/releaser/broadcaster for expression publication/distribution/release/broadcast date for expression copyright date for expression date of recording for expression numeric and/or alphabetic designations for serial expressions chronological designations for serial expressions performance date for expression place of performance for expression extent of expression content of expression language of expression text language of expression captions language of expression sound track language of sung or spoken text of expression language of expression subtitles language of expression intertitles language of summary or abstract of expression instrumentation of musical expression instrumentation of musical expression—number of a particular instrument instrumentation of musical expression—type of instrument voice(s) of musical expression voice(s) of musical expression—number of a particular type of voice voice(s) of musical expression—type of voice key of musical expression appendages to the expression expression series statement mode of issuance for expression notes about expression surrogate [under development] attributes/properties in rda compared to ycr (cont.) information technology and libraries | june model entity rda ycr manifestation title statement of responsibility edition statement numbering of serials production statement publication statement distribution statement manufacture statement copyright date series statement mode of issuance frequency identifier for the manifestation note media type carrier type base material applied material mount production method generation layout book format font size polarity reduction ratio sound characteristics projection characteristics of motion picture film video characteristics digital file characteristics equipment and system requirements terms of availability key identifier for manifestation publication statement of manifestation place of publication/distribution/release/broadcast of manifestation manifestation publisher/distributor/releaser/broadcaster manifestation date of publication/distribution/release/broadcast carrier edition statement carrier piece count carrier name carrier broadcast standard carrier recording type carrier playing speed carrier configuration of playback channels process used to produce carrier carrier dimensions carrier base materials carrier generation carrier polarity materials applied to carrier carrier encoding format intermediation tool requirements system requirements serial manifestation illustration statement manifestation standard number manifestation isbn manifestation issn manifestation publisher number manifestation universal product code notes about manifestation title- manifestation key identifier for title-manifestation variant title for title-manifestation title-manifestation title title-manifestation statement of responsibilities title-manifestation edition statement publication statement of title-manifestation place of publication/distribution/release/broadcasting of title- manifestation publisher/distributor/releaser, broadcaster of title-manifestation date of publication/distribution/release/broadcast of title- manifestation title-manifestation series title-manifestation mode of issuance notes about title-manifestation title-manifestation standard number attributes/properties in rda compared to ycr (cont.) can bibliographic data be put directly onto the semantic web? | yee model entity rda ycr serial title key identifier for serial title variant title for serial title title of serial title serial title statement of responsibility serial title edition statement publication statement of serial title place of publication/distribution/release/broadcast of serial title publisher/distributor/releaser/broadcaster of serial title date of publication/distribution/release/broadcast of serial title serial title beginning numeric and/or alphabetic designations serial title beginning chronological designations serial title ending numeric and/or alphabetic designations serial title ending chronological designations serial title frequency serial title mode of issuance serial title illustration statement notes about serial title serial title issn-l item preferred citation custodial history immediate source of acquisition identifier for the item item-specific carrier characteristics key identifier for item item barcode item location item call number or accession number item copy number item provenance item condition item marks and inscriptions item exhibition history item treatment history item scheduled treatment item access restrictions attributes/properties in rda compared to ycr (cont.) information technology and libraries | june model entity rda ycr person name of the person preferred name for the person variant name for the person date associated with the person title of the person fuller form of name other designation associated with the person gender place of birth place of death country associated with the person place of residence address of the person affiliation language of the person field of activity of the person profession or occupation biographical information identifier for the person key identifier for person language-based identifier (preferred lexical label) for person clan name of person forename/given name/first name of person matronymic of person middle name of person nickname of person patronymic of person surname/family name of person natural language order of forename, surname, middle name, patronymic, matronymic and/or clan name of person affiliation of person biography/history of person date of birth of person date of death of person ethnicity of person field of activity of person gender of person language of person place of birth of person place of death of person place of residence of person political affiliation of person profession/occupation of person religion of person variant name for person fictitious character [under development] performing animal [under development] corporate body name of the corporate body preferred name for the corporate body variant name for the corporate body place associated with the corporate body date associated with the corporate body associated institution other designation associated with the corporate body language of the corporate body address of the corporate body field of activity of the corporate body corporate history identifier for the corporate body key identifier for corporate body language-based identifier (preferred lexical label) for corporate body dates associated with corporate body field of activity of corporate body history of corporate body language of corporate body place associated with corporate body type of corporate body variant name for corporate body corporate subdivision [under development] place as jurisdictional corporate body [under development] attributes/properties in rda compared to ycr (cont.) can bibliographic data be put directly onto the semantic web? | yee model entity rda ycr conference or other event as corporate body creator [under development] jurisdictional corporate subdivision [under development] family name of the family preferred name for the family variant name for the family type of family date associated with the family place associated with the family prominent member of the family hereditary title family history identifier for the family concept term for the concept preferred term for the concept variant term for the concept type of concept identifier for the concept key identifier for concept language-based identifier (preferred lexical label) for concept qualifier for concept language-based identifier variant name for concept object name of the object preferred name for the object variant name for the object type of object date of production place of production producer/fabricator physical medium identifier for the object key identifier for object language-based identifier (preferred lexical label) for object qualifier for object language-based identifier variant name for object event name of the event preferred name for the event variant name for the event date associated with the event place associated with the event identifier for the event key identifier for event or historical period as subject language-based identifier (preferred lexical label) for event or historical period as subject beginning date for event or historical period as subject ending date for event or historical period as subject variant name for event or historical period as subject place name of the place preferred name for the place variant name for the place coordinates other geographical information identifier for the place key identifier for place as geographic area language-based identifier (preferred lexical label) for place as geographic area qualifier for place as geographic area variant name for place as geographic area discipline key identifier for discipline language-based identifier (preferred lexical label) (name or classification number or symbol) for discipline translation of meaning of classification number or symbol for discipline attributes/properties in rda compared to ycr (cont.) information technology and libraries | june model entity rda ycr genre/form key identifier for genre/form language-based identifier (preferred lexical label) for genre/form variant name for genre/form name scope of usage date of usage identifier controlled access point rules agency note: in rda, the following attributes have not yet been assigned to a particular class or entity: extent, dimensions, terms of availability, contact information, restrictions on access, restrictions on use, uniform resource locator, status of identification, source consulted, cataloguer’s note, status of identification, and undifferentiated name indicator. name is being treated as both a class and a property. identifier and controlled access point are treated as properties rather than classes in both rda and ycr. attributes/properties in rda compared to ycr (cont.) lauren h. mandel (lmandel@fsu.edu) is a doctoral candidate at the florida state university college of communication & informa- tion, school of library & information studies, and is research coordinator at the information use management & policy insti- tute. geographic information systems: tools for displaying in-library use data lauren h. mandel geographic information systems: tools for displaying in-library use data | mandel in-library use data is crucial for modern libraries to understand the full spectrum of patron use, including patron self-service activities, circulation, and reference statistics. rather than using tables and charts to display use data, a geographic information system (gis) facili- tates a more visually appealing graphical display of the data in the form of a map. giss have been used by library and information science (lis) researchers and practitio- ners to create maps that display analyses of service area populations and demographics, facilities space manage- ment issues, spatial distribution of in-library use of materials, planned branch consolidations, and so on. the “seating sweeps” method allows researchers and librari- ans to collect in-library use data regarding where patrons are locating themselves within the library and what they are doing at those locations, such as sitting and reading, studying in a group, or socializing. this paper proposes a gis as a tool to visually display in-library use data col- lected via “seating sweeps” of a library. by using a gis to store, manage, and display the data, researchers and librarians can create visually appealing maps that show areas of heavy use and evidence of the use and value of the library for a community. example maps are included to facilitate the reader’s understanding of the possibilities afforded by using giss in lis research. t he modern public library operates in a context of limited (and often continually reduced) funding where the librarians must justify the continued value of the library to funding and supervisory authori- ties. this is especially the case as more and more patrons access the library virtually, calling into question the relevance of the physical library. in this context, there is a great need for librarians and researchers to evaluate the use of library facility space to demonstrate that the physical library is still being used for important social and educational functions. despite this need, no model of public library facility evaluation emphasizes the ways patrons use library facilities. the systematic collection of in-library use data must go beyond traditional circula- tion and reference transactions to include self-service activities, group study and collaboration, socializing, and more. geographic information systems (giss) are beginning to become deployed in library and information science (lis) research as a tool for graphically displaying data. an initial review of the literature has yielded studies where a gis has been used in analyzing service area populations through u.s. census data; sitting facility locations; managing facilities, including spatial distribu- tion of in-library book use and occupancy of library study space; and planning branch consolidations. these uses of gis are not mutually exclusive; studies have combined multiple uses of giss. also, giss have been proposed as viable tools for producing visual representations of measurements of library facility use. these studies show the capabilities of a gis for storing, managing, analyzing, and displaying in-library use data and the value of gis- produced maps for library facility evaluations, in-library use research, and library justification. n research purpose observing and measuring the use of a library facility is a crucial step in the facility evaluation process. the library needs to understand how the facility is currently being used in order to justify the continued financial support necessary to maintain and operate it. understanding how the facility is used can also help librarians identify high- traffic areas of the library that are ideal locations to mar- ket library services and materials. this understanding cannot be reached by analyzing circulation and reference transaction data alone; it must include in-library use mea- sures that account for all ways patrons are using the facil- ity. the purpose of this paper is to suggest a method by which to observe and record all uses of a library facility during a sampling period, the so-called “seating sweep” performed by given and leckie, and then to use a gis to store, manage, and display the collected data on a map or series of maps that graphically depict library use. n significance of facility evaluation facility evaluation is a topic of vital importance in all fields, but this is especially true of a field such as public librarianship where funding is often a source of concern. in times of economic instability, libraries can benefit from the ability to identify uses of existing facilities and employ this information to justify the continued opera- tion of the library facility. also, knowing which areas of the library are more frequently used than others can help lauren h. mandel (lmandel@fsu.edu) is a doctoral candidate at the florida state university college of communication & in- formation, school of library & information studies, and is re- search coordinator at the information use management & policy institute. information technology and libraries | march librarians determine where to place displays of library materials and advertisements of library services. for a library to begin to evaluate patron use and how well the facility meets users’ needs, there must be an understanding of what users need from the library facility. to determine those needs, it is vital that library staff observe the facility while it is being used. this obser- vation can be applied to the facility evaluation plan to justify the continued operation of the facility to meet the needs of the library service population. understanding how people use the public library facility beyond traditional measures of circulation statis- tics and reference transactions can lead to new theories of library use, an area of significant research interest for lis. additionally, the importance of this work transcends lis because it applies to other government-funded com- munity service agencies as well. for example, recreation facilities and community centers could also benefit from a customer-use model that incorporates measures of the true use of those facilities. n literature review although much has been written on the use of library facilities, little of the research includes studies of how patrons actually use existing public library facilities and whether facilities are designed to accommodate this use. rather, much of the research in public library facility eval- uation has focused on collection and equipment space needs, despite the user-oriented focus of public library accountability models. recent research in library facility design is beginning to reflect this focus, but additional study would be useful to the field. use of gis is on the rise in the modern technologi- cal world. a gis is a computer-based tool for compiling, storing, analyzing, and displaying data graphically. usually this data is geospatial in nature, but a gis also can incorporate descriptive or statistic data to provide a richer picture than figures and tables can. although gis has been around for half a century, it has become increas- ingly more affordable, allowing libraries and similar institutions to consider using a gis as a measurement and analysis tool. giss have started being used in lis research as a tool for graphically displaying library data. one fruitful area has been the mapping of user demographics for facil- ity planning purposes, including studies that mapped library closures. mapping also can include in-library use data, in which case a gis is used to overlay collected in-library use data on library floor plans. this can offer a richer picture of how a facility is being used than tradi- tional charts and tables can provide. using a gis to display library service area population data adkins and sturges suggest libraries use a gis-based library service area assessment as a method to evaluate their service areas and plan library services to meet the unique demographic demands of their communities. they discuss the methods of using gis, including down- loading u.s. census tiger (topologically integrated geographic encoding and referencing) files, geocoding library locations, delineating service areas by multiple methods, and analyzing demographics. a key tenet of this approach is the concept that public libraries need to understand the needs of their patrons. this is a prevailing concept in the literature. prieser and wang, in reporting a method used to create a facilities master plan for the public library of cincinnati and hamilton county, ohio, offer a convincing argument for combining gis and building performance evaluation (bpe) methods to examine branch facility needs and offer individualized facilities recommendations. like other lis researchers, preiser and wang suggest a relation- ship between libraries and retail stores, noting the similar modern trends of destination libraries and destination bookstores. they also acknowledge the difficulty in com- pleting an accurate library performance assessment due to the multitude of activities and functions of a library. their method is a combination of a gis-based service area and population analysis with a bpe that includes staff and user interviews and surveys, direct observation, and photography. the described multimethod approach offers a more complete picture of a library facility’s per- formance than traditional circulation-based evaluations. further use of giss in library facility planning can be seen from a study comparing proposed branches by demographic data that has been analyzed and presented through a gis. hertel and sprague describe research that used a gis to conduct geospatial analysis of u.s. census data to depict the demographics of populations that would be served by two proposed branch libraries for a public library system in idaho. a primary purpose of this research is to demonstrate the possible ways public libraries can use gis to present visual and quantitative demographic analyses of service area populations. hertel and sprague identify that public libraries are challenged to determine which public they are serving and the needs of that population, writing that “libraries are beginning to add customer-based satisfaction as a critical compo- nent of resource allocation decisions” and need the help of a gis to provide hard-data evidence in support of staff observations. this evidence could take the form of demographic data, as discussed by hertel and sprague, and also could incorporate in-library use data to present a fuller picture of a facility’s use. geographic information systems: tools for displaying in-library use data | mandel using gis to display in-library use data xia conducted several studies in which he collected library- use data and mapped that data via a gis. in one study designed to identify the importance of space management in academic libraries, xia suggests applications of giss in library space management, particularly his tool integrating library floor plans with feature data in a gis. he explains that a gis can overcome the constraints of drafting and computer automated design tools, such as those in use at chico meriam library at california state university and at the michigan state university main library. for example, giss are not limited to space visualization manipulation, but can incorporate user perceptions, behavior, and daily activities, all of which are important data to library space management considerations and in-library use research. xia also reviews the use of gis tools that incorporate hos- pital and casino floor plans, noting that library facilities are as equally complex as hospitals and casinos; this is a com- pelling argument that academic libraries should consider the use of a gis as a space management tool. in another study, xia uses a gis to visualize the spatial distribution of books in the library in an attempt to establish the relationship between the height of book- shelves and the in-library use of books. this study seeks to answer the question of how the location of books on shelves of different heights could influence user behav- ior (i.e., patrons may prefer to browse shelves at eye level rather than the top and bottom shelves). what is of interest here is xia’s use of a gis to spatially represent the collected data. xia remarks that a gis “is suitable for assisting in the research of in-library book use where library floor layouts can be drawn into maps on multiple- dimensional views.” in fact, xia’s graphics depict the use of books by bookshelf height in a visual manner that could not be achieved without the use of a gis. similarly, a gis can be used to spatially represent the collected data in an in-library use study by overlaying the data onto a representation of the library floor plan. in a third project, xia measures study space use in academic libraries as a metric of user satisfaction with library services. he says that libraries need to evaluate space needs on case-by-case basis because every library is unique and serves a unique population. therefore, to observe the occupancy of study areas in an academic library, xia drew the library’s study facilities (including furniture) in a gis. he then observed patrons’ use of the facilities and entered the observation data into the gis to overlay on maps of the study areas. there are several advantages of using gis in this way: spatial databases can store continuing data sets, the system is powerful and flexible for manipulating and analyzing the spatial dataset, there are enhanced data visualization capabili- ties, and maps and data become interactive. conclusions drawn from the literature a gis is a tool gaining momentum in the world of lis research. giss have been used to conduct and display service area population assessments, propose facility locations, and plan for and measure branch consolidation impacts and benefits. giss also have been used to graphi- cally represent in-library use for managing facility space allocation, mapping in-library book use, and visualizing the occupancy of library study space. additionally, giss have been used in combination studies that examine library service areas and facility location proposals. these uses of giss are only the beginning; a gis can be used to map any type of data a library can collect, including all measures of in-library use. additionally, gis-based data analysis and display complements the focus in library-use research on gathering data to show a richer picture of a facility’s use and the focus in library facility design literature on build- ing libraries on the basis of community needs. n in-library use research that would benefit from spatial data displays unobtrusive observational research offers a rich method for identifying and recording the use of a public library facility. a researcher could obtain a copy of a library’s floor plan, predetermine sampling times during which to “sweep” the library, and conduct the sweeps by marking all patrons observed on the floor plan. this data then could be entered into a gis database for spatial analysis and display. specific questions that could be addressed via such a method include the following: n what are all the ways in which people are using the library facility? n how many people are using traditional library resources, such as books and computers? n how many people are using the facility for other rea- sons, such as relaxation, meeting friends, and so on? n do the ways in which patrons use the library vary by location within the facility (e.g., are the people using traditional library resources and the people using the library for other reasons using the same areas of the library or different areas)? n which area(s) of the library facility receive the highest level of use? it is hoped that answers to these questions, in whole or in part, could begin to offer a picture of how a library facility is currently being used by library patrons. to better view this picture, the data recorded from the observational research could be entered into a gis to information technology and libraries | march overlay onto the library floor plan in a similar manner as xia’s use of a gis to display occupancy of library study space. this spatial representation of the data should facilitate greater understanding of the actual use of the library facility. instead of a library presenting tables and graphs of library use, it would be able to produce illus- trative maps that would help explain patterns of use to funding and supervising authorities. these maps would not require expensive proprietary gis packages; the examples provided in this paper were created using the free, open-source mapwindow gis package. example using gis to display in-library use data for this paper, i produced example maps on the basis of fictional in-library use data. these maps were created using mapwindow gis software along with microsoft excel, publisher, and paint (see figure for a diagram of this process). mapwindow is an open-source gis package that is easy to learn and use, but its layout and graphic design features are limited compared to the more expensive and sophisticated proprietary gis packages. mapwindow files are compatible with the proprietary packages, so they could be imported into other gis packages for finishing. for this paper, however, the goal was to create simple maps that a novice could replicate. therefore publisher and paint were used for finalizing the maps, instead of a sophisticated gis package. it was relatively easy to create the maps. first, i drew a sample floor plan of a fictional library computer lab in excel and imported it into mapwindow as a jpeg file. i then overlaid polygons (shapes that represent area units such as chairs and tables) onto the floor plan and saved two shapefiles, one for tables and one for computers. a shapefile is a basic storage file used in most gis pack- ages. for each of those shapefiles i created an attribute table (basically, a linked spreadsheet) using fictitious data representing use of the tables and computers at and a.m. and , , , and p.m. on a sample day. the field cal- culator generated a final column summing the total use of each table and computer for the fictitious sample day. i then created maps depicting the use of both tables and computers at each of the sample time periods (see figure ) and for the total use (see figure ). benefits of gis-created displays for library managers the maps presented here are not based on actual data, but are meant to demonstrate the capabilities of giss for spa- tially representing the use of a library facility. this could be done on a grander scale using an entire library floor plan and data collected during a longer sample period (e.g., a full week). these maps can serve several purposes for figure . process diagram for creating the sample maps figure . example maps depicting use of tables and computers in a fictional library computer lab, by hour geographic information systems: tools for displaying in-library use data | mandel library managers, specifically regarding the marketing of library services and the justification of library funding. mapping data obtained from library “sweeps” can help identify the popularity of different areas of the library at different times of the day, different days of the week, or different times of the year. once the library has identified the most popular areas, this information can be used to market library materials and services. for example, a highly populated area would be an ideal loca- tion over which to install ceiling-mounted signs that the library could use for marketing services and programs. or the library could purchase a book display table similar to those used in bookstores and install it in the middle of a frequently populated area. the library could stock the table with seasonally relevant books and other materials (e.g., tax guidebooks in march and april) and track the circulation of these materials to determine the degree to which placement on the display table resulted in increased borrowing of those materials. in addition to helping the library market its materials and services, mapping in-library use can provide visual evidence of the library’s value. public libraries often rely on reference and circulation transaction data, gate counts, and programming attendance statistics to justify their exis- tence. these measures, although valuable and important, do not include many other ways that patrons use libraries, such as sitting and reading, studying, group work, and socializing. during “seating sweeps,” the observers can record any and all uses they observe, including any that may not have been anticipated. all of these uses could then be mapped, providing a richer picture of how a pub- lic library is used and stronger justification of the library’s value. these maps may be easier for funding and supervis- ing authorities to understand than textual explanations or graphs and charts of statistical analyses. n conclusion from a review of the literature, it is clear that giss are increasingly being used in lis research as data-analysis and display tools. giss are being used to analyze patron and materials data as well as studies combining com- bined multiple uses of giss. patron analysis has included service-area-population analysis and branch-consolida- tion planning. analysis of library materials has been used for space management, visualizing the spatial distribu- tion of in-library book use, and visual representation of facility-use measurements. this paper has proposed collecting in-library use data according to given and leckie’s “seating sweeps” method and visually displaying that data via a gis. examples of such visual displays were provided to facilitate the reader’s understanding of the possibilities afforded by using a gis in lis research, as well as the scalable nature of the method. librarians and library staff can produce maps similar to the examples in this paper with minimal gis training and background. the literature review and example figures offered in this paper show the capa- bilities of giss for analyzing and graphically presenting library-use data. giss are tools that can facilitate library facility evaluations, in-library use research, and library valuation and justification. references . denice adkins and denyse k. sturges, “library service planning with gis and census data,” public libraries , no. ( ): – ; karen hertel and nancy sprague, “gis and cen- sus data: tools for library planning,” library hi tech , no. ( ): – ; wolfgang f. e. preiser and xinhao wang, “assess- ing library performance with gis and building evaluation meth- ods,” new library world , no. – ( ): – . figure . example map depicting total use of tables and computers in a fictional library computer lab for a sample day information technology and libraries | march . hertel and sprague, “gis and census data”; preiser and wang, “assessing library performance.” . jingfeng xia, “library space management: a gis pro- posal,” library hi tech , no. ( ): – ; xia, “using gis to measure in-library book-use behavior,” information technology & libraries , no. ( ): – ; xia, “visualizing occupancy of library study space with gis maps,” new library world , no. – ( ): – . . preiser and wang, “assessing library performance.” . hertel and sprague, “gis and census data”; preiser and wang, “assessing library performance.” . preiser and wang, “assessing library performance”; xia, “library space management”; xia, “using gis to measure”; xia, “visualizing occupancy.” . lisa m. given and gloria j. leckie, “‘sweeping’ the library: mapping the social activity space of the public library,” library & information science research , no. ( ): – . . “jackson rejects levy to reopen libraries,” american libraries , no. ( ): – ; “may levy set for jackson county libraries closing in april,” american libraries , no. ( ): ; “tax reform has florida bracing for major budget cuts,” american libraries , no. ( ): . . anne morris and elizabeth barron, “user consultation in public library services,” library management , no. ( ): – ; susan l. silver and lisa t. nickel, surveying user activ- ity as a tool for space planning in an academic library (tampa: univ. of south florida library, ); james simon and kurt schlichting, “the college connection: using academic support to conduct public library services,” public libraries , no. ( ): – . . given and leckie, “‘sweeping’ the library”; christie m. koontz, dean k. jue, and keith curry lance, “collecting detailed in-library usage data in the u.s. public libraries: the methodology, the results and the impact,” in proceedings of the third northumbria international conference on performance measurement in libraries and information services (newcastle, uk: university of northumbria, ): – ; koontz, jue, and lance, “neighborhood-based in-library use performance measures for public libraries: a nationwide study of majority- minority and majority white/low income markets using personal digital data collectors,” library & information science research , no. ( ): – . . cheryl bryan, managing facilities for results: optimizing space for services (chicago: ala, ); anders c. dahlgren, public library space needs: a planning outline (madison, wis.: department of public instruction, ); william w. sannwald and robert s. smith, eds., checklist of library building design considerations (chicago: ala, ). . brenda dervin, “useful theory for librarianship: com- munication, not information,” drexel library quarterly , no. ( ): – ; morris and barron, “user consultation”; pre- iser and wang, “assessing library performance”; simon and schlichting, “the college connection”; norman walzer, karen stott, and lori sutton, “changes in public library services,” illinois libraries , no. ( ): – . . bradley wade bishop, “use of geographic information systems in marketing and facility site location: a case study of douglas county (colo.) libraries,” public libraries , no. : – ; david jones, “people places: public library build- ings for the new millennium,” australasian public libraries & information services , no. ( ): – ; nolan lushington, libraries designed for users: a st century guide (new york: neal-schuman, ); shannon mattern, “form for function: the architecture for new libraries,” in the new downtown library: designing with communities (minneapolis: univ. of minnesota pr., ), – . . united nations, department of economic and social affairs, statistics division, handbook on geographical information systems and mapping (new york: united nations, ). . adkins and sturges, “library service planning”; bishop, “use of geographic information systems”; hertel and sprague, “gis and census data”; christie koontz, “using geographic information systems for estimating and profiling geographic library market areas,” in geographic information systems and libraries: patrons, maps, and spatial information, ed. linda c. smith and mike gluck (urbana–champaign: univ. of illinois pr., ): – ; preiser and wang, “assessing library perfor- mance.” . christie m. koontz, dean k. jue, and bradley wade bishop, “public library facility closure: an investigation of reasons for closure and effects on geographic market areas,” library & information science research , no. ( ): – . . xia, “library space management”; xia, “using gis to measure”; xia, “visualizing occupancy.” . adkins and sturges, “library service planning.” . bishop, “use of geographic information systems”; jones, “people places”; koontz, jue, and lance, “collecting detailed in- library usage data”; koontz, jue, and lance, “neighborhood- based in-library use”; morris and barron, “user consultation”; simon and schlichting, “the college connection”; walzer, stott, and sutton, “changes in public library services.” . preiser and wang, “assessing library performance.” . given and leckie, “‘sweeping’ the library;” christie m. koontz, “retail interior layout for libraries,” marketing library services , no. ( ): – . . hertel and sprague, “gis and census data.” . ibid., . . xia, “library space management.” . xia, “using gis to measure.” . ibid., . . xia, “visualizing occupancy.” . adkins and sturges, “library service planning”; her- tel and sprague, “gis and census data”; preiser and wang, “assessing library performance.” . hertel and sprague, “gis and census data”; preiser and wang, “assessing library performance.” . koontz, jue, and bishop, “public library facility clo- sure”; preiser and wang, “assessing library performance.” . xia, “library space management”; xia, “using gis to measure”; xia, “visualizing occupancy.” . hertel and sprague, “gis and census data”; preiser and wang, “assessing library performance.” . given and leckie, “‘sweeping’ the library”; koontz, jue, and lance, “collecting detailed in-library usage data”; koontz, jue, and lance, “neighborhood-based in-library use”; silver and nickel, surveying user activity; jones, “people places”; lushington, libraries designed for users. . given and leckie, “‘sweeping’ the library.” . xia, “visualizing occupancy.” . for more information or to download mapwindow gis, see http://www.mapwindow.org/ editorial | truitt marc truitteditorial w elcome to ! it has been unseasonably cold in edmonton, with daytime “highs”—i use the term loosely— averaging around - °c (that’s - °f, for those of you ital readers living in the states) for much of the last three weeks. factor in wind chill (a given on the canadian prairies), and you can easily subtract another °c. as a result, we’ve had more than a few days and nights where the adjusted temperature has been much closer to - °, which is the same in either celsius or fahrenheit. while my boss and chief librarian is fond of saying that “real canadians don’t even button their shirts until it gets to minus forty,” i’ve yet to observe such a feat of derring-do by anyone at much less than twenty below <grin>. even your editor’s two labrador retrievers—who love cooler weather—are reluctant to go out in such cold, with the result that both humans and pets have all been coping with bouts of cabin fever since before christmas. n so, when is it “too cold” for a server room? why, you may reasonably ask, am i belaboring ital readers with the details of our weather? over the week- end we experienced near-simultaneous failures of both cooling systems in our primary server room (sr ), which meant that nearly all of our library it services, including our opac (which we host for a consortium of twenty area libraries), a separate opac for edmonton public library, our website, and access to licensed e-resources, e-mail, files, and print servers had to be shut down. temperature readings in the room soared from an average of – °c ( – . °f) to as much as °c ( . °f) before settling out at around °c ( °f). we spent much of the weekend and beginning of this week relocating servers to all man- ner of places while the cooling system gets fixed. i imag- ine that next we may move one into each staff person’s under-heated office, where they’ll be able to perform double duty as high-tech foot warmers! all of this happened, of course, while the temperature outside the building hovered between - ° and - °c. this is not the first time we’ve experienced a failure of our cooling systems during extremely cold weather. last winter we suffered a series of problems with both the systems in sr and in our secondary room a few feet away. the issues we had then were not the same as those we’re living through now, but they occurred, as now, at the coldest time of the year. this seeming dichotomy of an overheated server environment in the depths of winter is not a matter of accident or coincidence; indeed, while it may seem counterintuitive, the fact is that many, if not all, of our cooling woes can be traced to the cold outside. the simple explanation is that extreme cold weather stresses and breaks things, including hvac systems. as we’ve tried to analyze this incident, it appears likely that our troubles began when the older of our two systems in sr developed a coolant leak at some point after its last preventive maintenance servicing in august. fall was mild here, and we didn’t see the onset of really severe cold weather until early to mid-december. since the older system is mainly intended for failover of the newer one, and since both systems last received routine service recently, it is possible that the leak could have developed at any time since, although my supposition is that it may be itself a result of the cold. in any case, all seemed well because the newer cool- ing system in sr was adequate to mask the failure of the older unit, until it suffered a controller board failure that took it offline last weekend. but, with the failure of the new system on saturday, all it services provided from this room had to be brought down. after a night spent try- ing to cool the room with fans and a portable cooling unit, we succeeded in bringing the two opacs and other core services back online by sunday, but the coolant leak in the old system was not repaired until midday monday. today is friday, and we’ve limped along all week on about percent of the cooling normally required in sr . we hope to have the parts to repair the newer cooling system early next week (fingers crossed!). some interesting lessons have emerged from this incident, and while probably not many of you regularly deal with - °c winters, i think them worth sharing in the hope that they are more generally applicable than our winter extremes are: . document your servers and the services that reside on them. we spent entirely too much time in the early hours of this event trying to relate servers and ser- vices. we in information technology (it) may think of shutting down or powering up servers “fred,” “wilma,” “betty,” and “barney,” but, in a crisis, what we generally should be thinking of is whether or not we can shut down e-mail, file-and-print ser- vices, or the integrated library system (ils) (and, if the latter, whether we shut down just the underlying database server or also the related staff and public services). perhaps your servers have more obvious names than ours, in which case, count yourself for- tunate. but ours are not so intuitively named—there is a perfectly good reason for this, by the way—and with distributed applications where the database marc truitt (marc.truitt@ualberta.ca) is associate director, bibliographic and information technology services, university of alberta libraries, edmonton, alberta, canada, and editor of ital. information technology and libraries | march may reside here, the application there, and the web front end yet somewhere else, i’d be surprised if your situation isn’t as complex as ours. and bear in mind that documentation of dependencies goes two ways: not only do you want to know that “barney” is hosting the ils’s oracle database, but you also want to know all of the servers that should be brought up for you to offer ils–related services. . prioritize your services. if your cooling system (or other critical server-room utility) were sud- denly only operating at percent of your normal required capacity, how would you quickly decide which services to shut down and which to leave up? i wrote in this space recently that we’ve been thinking about prioritized services in the context of disaster recovery and business continuity, but this week’s incident tells me that we’re not really there yet. optimally, i think that any senior member of my on-call staff should be empowered in a given critical situation to bring down services on the basis of a predefined set of service priorities. . virtualize, virtualize, virtualize. if we are at all typi- cal of large libraries in the association of research libraries (and i think we are), then it will come as no surprise that we seem to add new services with alarming frequency. i suspect that, as with most places, we tend to try and keep things simple at the server end by hosting new services on sepa- rate, dedicated servers. the resulting proliferation of new servers has led to ever-greater strains on power, cooling, and network infrastructures in a facility that was significantly renovated less than two years ago. and i don’t see any near-term likeli- hood that this will change. we are, consequently, in the very early days of investigating virtualization technology as a means of reducing the number of physical boxes and making much better use of the resources—especially processor and ram— available to current-generation hardware. i’m hop- ing that someone among our readership is farther along this path than we and will consider submit- ting to ital a “how we done it” on virtualization in the library server room very soon! . sometimes low-tech solutions work . . . no one here has failed to observe the irony of an overheated server room when the temperature just steps away is ° below. our first thought was how simple and elegant a solution it would be to install duct- ing, an intake fan, and a damper to the outside of the building. then, the next time our cooling failed in the depths of winter, voila!, we could solve the problem with a mere turn of the damper control. . . . . and sometimes they don’t. not quite, it seems. when asked, our university facilities experts told us that an even greater irony than the one we currently have would be the requirement for can$ , in equipment to heat that - °c outside air to around freezing so that we wouldn’t freeze pipes and other indoor essentials if we were to adopt the “low-tech” approach and rely on mother nature. oh, well . . . n in memoriam most of the snail mail i receive as editor consists of advertisements and press releases from various firms providing it and other services to libraries. but a few months ago a thin, hand-addressed envelope, post- marked pittsburgh with no return address, landed on my desk. inside were two slips of paper clipped from a recent issue of ital and taped together. on one was my name and address; the other was a mailing label for jean a. guasco of pittsburgh, an ala life member and ital subscriber. beside her name, in red felt-tip pen, someone had written simply “deceased.” i wondered about this for some time. who was ms. guasco? where had she worked, and when? had she published or otherwise been active professionally? if she was a life member of ala, surely it would be easy to find out more. it turns out that such is not the case, the wonders of the internet notwithstanding. my obvious first stop, google, yielded little other than a brief notice of her death in a pittsburgh-area newspaper and an entry from a digi- tized september issue of special libraries that identi- fied her committee assignment in the special libraries assocation and the fact that she was at the time the chief librarian at mcgraw-hill, then located in new york. as a result of checking worldcat, where i found a listing for her master’s thesis, i learned that she graduated from the now-closed school of library service at columbia university in . if she published further, there was no mention of it on google. my subsequent searches under her name in the standard online lis indexes drew blanks. from there, the trail got even colder. mcgraw-hill long ago forsook new york for the wilds of ohio, and it seems that we as a profession have not been very good at retaining for posterity our directories of those in the field. a friend managed to find listings in both the – and – volumes of who’s who in special libraries, but all these did was confirm what i already knew: ms. guasco was an ala life member, who by then lived in pittsburgh. i’m guessing that she was then retired, since her death notice gave her age as eighty-six years. of her professional career before that, i’m sad that i must say i was able to learn no more. a semantic model of selective dissemination of information | morales-del-castillo et al. a semantic model of selective dissemination of information for digital libraries j. m. morales-del-castillo, r. pedraza-jiménez, a. a. ruíz, e. peis, and e. herrera-viedma in this paper we present the theoretical and methodo- logical foundations for the development of a multi-agent selective dissemination of information (sdi) service model that applies semantic web technologies for spe- cialized digital libraries. these technologies make pos- sible achieving more efficient information management, improving agent–user communication processes, and facilitating accurate access to relevant resources. other tools used are fuzzy linguistic modelling techniques (which make possible easing the interaction between users and system) and natural language processing (nlp) techniques for semiautomatic thesaurus genera- tion. also, rss feeds are used as “current awareness bul- letins” to generate personalized bibliographic alerts. n owadays, one of the main challenges faced by information systems at libraries or on the web is to efficiently manage the large number of docu- ments they hold. information systems make it easier to give users access to relevant resources that satisfy their information needs, but a problem emerges when the user has a high degree of specialization and requires very specific resources, as in the case of researchers. in “tra- ditional” physical libraries, several procedures have been proposed to try to mitigate this issue, including the selec- tive dissemination of information (sdi) service model that make it possible to offer users potentially interesting documents by accessing users’ personal profiles kept by the library. nevertheless, the progressive incorporation of new information and communication technologies (icts) to information services, the widespread use of the internet, and the diversification of resources that can be accessed through the web has led libraries through a process of reinvention and transformation to become “digital” libraries. this reengineering process requires a deep revision of work techniques and methods so librarians can adapt to the new work environment and improve the services provided. in this paper we present a recommendation and sdi model, implemented as a service of a specialized digital library (in this case, specialized in library and informa- tion science), that can increase the accuracy of accessing information and the satisfaction of users’ information needs on the web. this model is built on a multi-agent framework, similar to the one proposed by herrera-viedma, peis, and morales-del-castillo, that applies semantic web technologies within the specific domain of special- ized digital libraries in order to achieve more efficient information management (by semantically enriching dif- ferent elements of the system) and improved agent–agent and user–agent communication processes. furthermore, the model uses fuzzy linguistic model- ling techniques to facilitate the user–system interaction and to allow a higher grade of automation in certain procedures. to increase improved automation, some natural language processing (nlp) techniques are used to create a system thesaurus and other auxiliary tools for the definition of formal representations of information resources. in the next section, “instrumental basis,” we briefly analyze sdi services and several techniques involved in the semantic web project, and we describe the prelimi- nary methodological and instrumental bases that we used for developing the model, such as fuzzy linguistic model- ling techniques and tools for nlp. in “semantic sdi serv- ice model for digital libraries,” the bulk of this work, the application model that we propose is presented. finally, to sum up, some conclusive data are highlighted. n instrumental basis filtering techniques for sdi services filtering and recommendation services are based on the application of different process-management techniques that are oriented toward providing the user exactly the information that meets his or her needs or can be of his or her interest. in textual domains, these services are usu- ally developed using multi-agent systems, whose main aims are n to evaluate and filter resources normally repre- sented in xml or html format; and n to assist people in the process of searching for and retrieving resources. j. m. morales-del-castillo (josemdc@ugr.es) is assistant professor of information science, library and information science department, university of granada, spain. r. pedraza- jiménez (rafael.pedraza@upf.edu) is assistant professor of information science, journalism and audiovisual communication department, pompeu fabra university, barcelona, spain. a. a. ruíz (aangel@ugr.es) is full professor of information science, library and information science department, university of granada. e. peis (epeis@ugr.es) is full professor of information science, library and information science department, university of granada. e. herrera-viedma (viedma@decsai.ugr.es) is senior lecturer in computer science, computer science and artificial intelligence department, university of granada. information technology and libraries | march traditionally, these systems are classified as either content-based recommendation systems or collaborative recommendation systems. content-based recommen- dation systems filter information and generate recom- mendations by comparing a set of keywords defined by the user with the terms used to represent the content of documents, ignoring any information given by other users. by contrast, collaborative filtering systems use the information provided by several users to recommend documents to a given user, ignoring the representation of a document’s content. it is common to group users into different categories or stereotypes that are characterized by a series of rules and preferences, defined by default, that represent the information needs and common behav- ioural habits of a group of related users. the current trend is to develop hybrids that make the most of content-based and collaborative recommendation systems. in the field of libraries, these services usually adopt the form of sdi services that, depending on the profile of subscribed users, periodically (or when required by the user) generate a series of information alerts that describe the resources in the library that fit a user’s interests. sdi services have been studied in different research areas, such as the multi-agent systems development domain, and, of course, the digital libraries domain. presently, many sdi services are implemented on web platforms based on a multi-agent architecture where there is a set of intermediate agents that compare users’ profiles with the documents, and there are input-output agents that deal with subscriptions to the service and display generated alerts to users. usually, the information is struc- tured according to a certain data model, and users’ profiles are defined using a series of keywords that are compared to descriptors or the full text of the documents. despite their usefulness, these services have some deficiencies: n the communication processes between agents, and between agents and users, are hindered by the dif- ferent ways in which information is represented. n this heterogeneity in the representation of infor- mation makes it impossible to reuse such informa- tion in other processes or applications. a possible solution to these deficiencies consists of enriching the information representation using a common vocabulary and data model that are understandable by humans as well as by software agents. the semantic web project takes this idea and provides the means to develop a universal platform for the exchange of information. semantic web technologies the semantic web project tries to extend the model of the present web by using a series of standard languages that enable enriching the description of web resources and make them semantically accessible. to do that, the project basis itself on two fundamental ideas: ( ) resources should be tagged semantically so that informa- tion can be understood both by humans and comput- ers, and ( ) intelligent agents should be developed that are capable of operating at a semantic level with those resources and that infer new knowledge from them (shift- ing from the search of keywords in a text to the retrieval of concepts). the semantic backbone of the project is the resource description framework (rdf) vocabulary, which pro- vides a data model to represent, exchange, link, add, and reuse structured metadata of distributed information sources, thereby making them directly understandable by software agents. rdf structures the information into individual assertions (e.g., “resource,” “property,” and “property value triples”) and uniquely character- izes resources by means of uniform resource identifiers (uris), allowing agents to make inferences about them using web ontologies or other, simpler semantic struc- tures, such as conceptual schemes or thesauri. even though the adoption of the semantic web and its application to systems like digital libraries is not free from trouble (because of the nature of the technologies involved in the project and because of the project’s ambi- tious objectives, among other reasons), the way these technologies represent the information is a significant improvement over the quality of the resources retrieved by search engines, and it also allows the preservation of platform independence, thus favouring the exchange and reuse of contents. as we can see, the semantic web works with infor- mation written in natural language that is structured in a way that can be interpreted by machines. for this reason, it is usually difficult to deal with problems that require operating with linguistic information that has a certain degree of uncertainty (e.g., when quantifying the user’s satisfaction in relation to a product or service). a possible solution could be the use of fuzzy linguistic modelling techniques as a tool for improving system–user commu- nication. fuzzy linguistic modelling fuzzy linguistic modelling supplies a set of approxi- mate techniques appropriate for dealing with qualitative aspects of problems. the ordinal linguistic approach is defined according to a finite set of tags (s) completely ordered and with odd cardinality (seven or nine tags): { }{ }t,=hi,s=s i …∈ , the central term has a value of approximately . , and the rest of the terms are arranged symmetrically around a semantic model of selective dissemination of information | morales-del-castillo et al. it. the semantics of each linguistic term is given by the ordered structure of the set of terms, considering that each linguistic term of the pair (si, st-i) is equally informative. each label si is assigned a fuzzy value defined in the inter- val [ , ] that is described by a linear trapezoidal property function represented by the -tupla (ai, bi, αi, βi). (the two first parameters show the interval where the property value is . ; the third and fourth parameters show the left and right limits of the distribution.) additionally, we need to define the following properties: .–the set is ordered: si ≥ sj if i ≥ j. .–there is the negation operator: neg(si ) = sj, with j = t - i. .–maximization operator: max(si, sj) = si if si ≥ sj. .–minimization operator: min(si, sj) = si if si ≤ sj. it also is necessary to define aggregation operators, such as linguistic weighted averaging (lwa), capable of and operating with and combining linguistic information. focusing on facilitating the interaction between users and system, the other starting objective is to achieve the development and implementation of the model proposed in the most automated way possible. to do this, we use a basic auxiliary tool—a thesaurus—that, among other tasks, assists users in the creation of their profile and ena- bles automating the alerts generation. that is why it is critical to define the way in which we create this tool, and in this work we propose a specific method for the semiautomatic development of thesauri using nlp techniques. nlp techniques and other automating tools nlp consists of a series of linguistic techniques, statistic approaches, and machine learning algorithms (mainly clustering techniques) that can be used, for example, to summarize texts in an automatic way, to develop automatic translators, and to create voice recognition software. another possible application of nlp would be the semiautomatic construction of thesauri using different techniques. one of them consists of determining the lexical relations between the terms of a text (mainly syn- onymy, hyponymy, and hyperonymy), and extracting terms that are more representative for the text’s specific domain. it is possible to elicit these relations by using linguistic tools, like princeton’s wordnet (http://wordnet .princeton.edu) and clustering techniques. wordnet is a powerful multilanguage lexical data- base where each one of its entries is defined, among other elements, by their synonyms (synsets), hyponyms, and hyperonyms. as a consequence, once given the most important terms of a domain, wordnet can be used to create from them a thesaurus (after leaving out all terms that have not been identified as belonging or related to the domain of interest). this tool can also be used with clustering tech- niques—for example, to group documents of a collection in a set of nodes or clusters, depending on their similarity. each of these clusters is described by the most representa- tive terms of their documents. these terms make up the most specific level of a thesaurus and are used to search in wordnet for their synonyms and most general terms, contributing (with the repetition of this procedure) to the bottom-up-development process of the thesaurus. although there are many others, these are some of the most well-known techniques of semiautomatic thesau- rus generation (semiautomatic because, needless to say, the supervision of experts is necessary to determine the validity of the final result). for specialized digital libraries, we propose develop- ing, on a multi-agent platform and using all these tools, sdi services capable of generating alerts and recommendations for users according to their personal profiles. in particular, the model presented here is the result of several previous models merging, and its service is based on the definition of “current-awareness bulletins,” where users can find a basic description of the resources recently acquired by the library or those that might be of interest to them. n the semantic sdi service model for digital libraries the sdi service includes two agents (an interface agent and a task agent) distributed in a four-level hierarchi- cal architecture: user level, interface level, task level and resource level. its main components are a repository of full-text doc- uments (which make up the stock of the digital library) and a series of elements described using different rdf- based vocabularies: one or several rss feeds that play a role similar to that of current-awareness bulletins in traditional libraries; a repository of recommendation log files that store the recommendations made by users about the resources, and a thesaurus that lists and hierarchi- cally relates the most relevant terms of the specialization domain of the library. also, the semantics of each ele- ment (that is, its characteristics and the relations the ele- ment establishes with other elements in the system) are defined in a web ontology developed in web ontology language (owl). next, we describe these main elements as well as the different functional modules that the system uses to carry out its activity. elements of the model there are four basic elements that make up the system: information technology and libraries | march the thesaurus, user profiles, rss feeds, and recommenda- tion log files. thesaurus an essential element of this sdi service is the thesau- rus, an extensible tool used in traditional libraries that enables organizing the most relevant concepts in a specific domain, defining the semantic relations estab- lished between them, such as equivalence, hierarchical, and associative relations. the functions defined for the thesaurus in our system include helping in the indexing of rss feeds items and in the generation of information alerts and recommendations. to create the thesaurus, we followed the method suggested by pedraza-jiménez, valverde-albacete, and navia-vázquez. the learning technique used for the creation of a the- saurus includes four phases: preprocessing of documents, parameterizing the selected terms, conceptualizing their lexical stems, and generating a lattice or graph that shows the relation between the identified concepts. essentially, the aim of the preprocessing phase is to prepare the documents’ parameterization by removing elements regarded as superfluous. we have developed this phase in three stages: eliminating tags (stripping), standardizing, and stemming. in the first stage, all the tags (html, xml, etc.) that can appear in the collection of documents are eliminated. the second stage is the standardization of the words in the documents in order to facilitate and improve the parameterization process. at this stage, the acronyms and n-grams (bigrams and trigrams) that appear in the documents are identified using lists that were created for that purpose. once we have detected the acronyms and n-grams, the rest of the text is standardized. dates and numeri- cal quantities are standardized, being substituted with a script that identifies them. all the terms (except acro- nyms) are changed to small letters, and punctuation marks are removed. finally, a list of function words is used to eliminate from the texts articles, determiners, auxiliary verbs, conjunctions, prepositions, pronouns, interjections, contractions, and grade adverbs. all the terms are stemmed to facilitate the search of the final terms and to improve their calculation during parameterization. to carry out this task, we have used morphy, the stemming algorithm used by wordnet. this algorithm implements a group of functions that check whether a term is an exception that does not need to be stemmed and then convert words that are not exceptions to their basic lexical form. those terms that appear in the documents but are not identified by morphy are elimi- nated from our experiment. the parameterization phase has a minimum complex- ity. once identified, the final terms (roots or bases) are quantified by being assigned a weight. such weight is obtained by the application of the scheme term frequency- inverse document frequency (tf-idf), a statistic measure that makes possible the quantification of the importance of a term or n-gram in a document depending on its fre- quency of appearance and in the collection the document belongs to. finally, once the documents have been parameter- ized, the associated meanings of each term (lemma) are extracted by searching for them in wordnet (specifically, we use wordnet . for unix-like systems). thus we get the group of synsets associated with each word. the group of hyperonyms and hyponyms also are extracted from the vocabulary of the analyzed collection of documents. the generation of our thesaurus—that is, the identifi- cation of descriptors that better represent the content of documents, and the identification of the underlying rela- tions between them—is achieved using formal concept analysis techniques. this categorization technique uses the theory of lat- tices and ordered sets to find abstraction relations from the groups it generates. furthermore, this technique ena- bles clustering the documents depending on the terms (and synonyms) it contains. also, a lattice graph is gener- ated according to the underlying relations between the terms of the collection, taking into account the hypero- nyms and hyponyms extracted. in that graph, each node represents a descriptor (namely, a group of synonym terms) and clusters the set of documents that contain it, linking them to those with which it has any relation (of hyponymy or hyperonymy). once the thesaurus is obtained by identifying its terms and the underlying relations between them, it is automatically represented using the simple knowledge organization system (skos) vocabulary (see figure ). user profiles user profiles can be defined as structured representations that contain personal data, interests, and preferences of users with which agents can operate to customize the sdi service. in the model proposed here, these profiles are basically defined with friend of a friend (foaf), a specific rdf/xml for describing people (which favours the profile interoperability, since this is a widespread vocabulary supported by an owl ontology) and another nonstandard vocabulary of our own to define fields not included in foaf (see figure ). profiles are generated the moment the user is regis- tered in the system, and they are structured in two parts: a public profile that includes data related to the user’s identity and affiliation, and a private profile that includes the user’s interests and preferences about the topic of the alerts he or she wishes to receive. to define their preferences, users must specify key- words and concepts that best define their information a semantic model of selective dissemination of information | morales-del-castillo et al. needs. later, the system compares those concepts with the terms in the thesaurus using as a similarity measure the edit tree algorithm. this function matches character strings, then returns the term introduced (if there’s an exact match) or the lexically most similar term (if not). consequently, if the suggested term satisfies user expectations, it will be added to the user’s profile together with its synonyms (if any). in those cases where the suggested term is not satisfactory, the system must have any tool or application that enables users to browse the thesaurus and select terms that bet- ter describe their needs. an exam- ple of this type of applications is thmanager (http://thmanager .sourceforge.net), a project of the universidad de zaragoza, spain, that enables editing, visualiz- ing, and going through structures defined in skos. each of the terms selected by the user to define his or her areas of interest has an associated lin- guistic frequency value (tagged as <freq>) that we call “satisfaction frequency.” it represents the regular- ity with which a particular prefer- ence value has been used in alerts positively evaluated by the user. this frequency measures the relative importance of the preferences stated by the user and allows the interface agent to generate a ranking list of results. the range of possible values for these frequencies is defined by a group of seven labels that we get from the fuzzy linguistic variable “frequency,” whose expression domain is defined by the linguis- tic term set s = {always, almost_ always, often, occasionally, rarely, almost_never, never}, being the default value and “occasionally” being the central value. rss feeds thanks to the popularization of blogs, there has been wide- spread use of several vocabular- ies specifically designed for the syndication of contents (that is, for making accessible to other internet users the content of a website by means of hyperlink lists called “feeds”). to create our current-awareness bulletin we use rss . , a vocabulary that enables managing hyperlinks lists in an easy and flexible way. it utilizes the rdf/xml syntax and data model and is easily extensible because of the use of <skos:concept rdf:about=” ”> <skos:inscheme rdf:resource=”http://www.ugr.es/…/thes/”/> <skos:preflabel xml:lang=”es”>proceedings</skos:preflabel> <skos:broader rdf:resource=”http://www.ugr.es/…/thes/ ”/> <skos:narrower rdf:resource=”http://www.ugr.es/…/thes/ ”/> <skos:narrower rdf:resource=”http://www.ugr.es/…/thes/ ”/> </skos:concept> figure . sample entry of a skos core thesaurus <foaf:personalprofiledocument rdf:about=””> <foaf:maker rdf:resource=”#person”/> <foaf:primarytopic rdf:resource=”#person”/> </foaf:personalprofiledocument> <foaf:person rdf:id=”user_ ”> <foaf:name>diego allione</foaf:name> <foaf:title>sr.</foaf:title> <foaf:mbox_sha sum>af fa df e </foaf:mbox_sha sum> <foaf:homepage rdf:resource=”http://allione.org”/> <foaf:depiction rdf:resource=”allione.jpg”/> <foaf:phone rdf:resource=”tel: - - ”/> <dfss:topic> <dfss:pref rdf:nodeid=”pref_ - ”> <rdfs:label>library management</rdfs:label> <dfss:relev> . </dfss:relev> </dfss:pref> </dfss:topic> </foaf:person> figure . user profile sample information technology and libraries | march modules that enable extending the vocabulary without modi- fying its core each time new describing elements are added. in this model several modules are used: the dublin core (dc) module to define the basic bib- liographic information of the items utilizing the elements established by the dublin core metadata initiative (http:// dublincore.org); the syndica- tion module to facilitate soft- ware agents synchronizing and updating rss feeds; and the taxonomy module to assign topics to feeds items. the structure of the feeds comprises two areas: one where the channel itself is described by a series of basic metadata like a title, a brief description of the content, and the updating frequency; and another where the descriptions of the items that make up the feed (see figure ) are defined (including elements such as title, author, sum- mary, hyperlink to the primary resource, date of creation, and subjects). recommendation log file each document in the repository has an associated recommendation log file in rdf that includes the listing of evaluations assigned to that resource by different users since the resource was added to the system. each of the entries of the recom- mendation log files consists of a recommendation value, a uri that identifies the user that has done the recommendation, and the date of the record (see figure ). the expression domain of the rec- ommendations is defined by the following set of five fuzzy linguistic labels that are extracted from the linguistic variable “quality of the resource”: q = {very_low, low, medium, high, very_high}. these elements represent the raw materials for the sdi service that enable it to develop its activity through four processes or functional modules: the pro- files updating process, rss feeds generation process, alert generation process, and collaborative recommen- dation process. system processes profiles updating process since the sdi service’s functions are based on generating passive searches to rss feeds from the preferences stored <recomm-log rdf:id=”log- ”> <doc rdf:resource=”http://doc.es/doc- a ”/> <items_e> <item rdf:nodeid=”item- a ”> <user rdf:resource=”http://user.es/ ”/> <date> / / </date> <recomm>high</recomm> </item> </ítems_e> </recomm-log> figure . recommendation log file sample <item rdf:about=”http://www.ugr.es/…/doc- ”> <dc:creator>escudero sánchez, manuel</dc:creator> <dc:creator>fernández cáceres, josé luis</dc:creator> <title>broadcasting and the internet http://eprints.rclis.org/…/audiovideo_good.pdf this paper is about… redoc, ( ), virual communities figure . rss feed item sample in a user’s profile, updating the profiles becomes a critical task. user profiles are meant to store long-term prefer- ences, but the system must be able to detect any subtle change in these preferences over time to offer accurate recommendations. in our model, user profiles are updated using a simple mechanism that enables finding users’ implicit preferences by applying fuzzy linguistic techniques and taking into account the feedback users provide. users are asked about their satisfaction degree (ej) in relation to the informa- tion alert generated by the system (i.e., whether the items a semantic model of selective dissemination of information | morales-del-castillo et al. retrieved are interesting or not). this satisfaction degree is obtained from the linguistic variable “satisfaction,” whose expression domain is the set of five linguistic labels: s’ = {total, very_high, high, medium, low, very_low, null}. this mechanism updates the satisfaction frequency associated with each user preference according to the satisfaction degree ej. it requires the use of a matching function similar to those used to model threshold weights in weighted search queries. the function proposed here rewards the frequencies associated with the preference val- ues present when resources assessed are satisfactory, and it penalizes them when this assessment is negative. let ej { }t,=hba,|ss,s ba ,...∈∈ s’ be the degree of satisfaction, and f j i l { }t,=hba,|ss,s ba ,...∈∈ s the frequency of property i (in this case i = “preference”) with value l, then we define the updating function g as s’x s→s: { } { } ( ) {=f,eg s ‘http://www.doaj.org/oai.article’; # the oai repository mylibrary::config->instance( ‘articles’ ); # the mylibrary instance # create a facet called formats $facet = mylibrary::facet->new; $facet->facet_name( ‘formats’ ); $facet->facet_note( ‘types of physical items embodying information.’ ); $facet->commit; $formatid = $facet->facet_id; # create an associated term called articles $term = mylibrary::term->new; $term->term_name( ‘articles’ ); $term->term_note( ‘short, scholarly essays.’ ); $term->facet_id( $formatid ); $term->commit; $articleid = $term->term_id; # create a location type called url $location_type = mylibrary::resource::location::type->new; $location_type->name( ‘url’ ); $location_type->description( ‘the location of an internet resource.’ ); $location_type->commit; $location_type_id = $location_type->location_type_id; # create a harvester and loop through each oai set mylibrary: a digital library framework and toolkit | morgan $harvester = net::oai::harvester->new( ‘baseurl’ => doaj ); $sets = $harvester->listsets; foreach ( $sets->setspecs ) { # get each record in this set and process it $records = $harvester->listallrecords( metadataprefix => ‘oai_dc’, set => $_ ); while ( $record = $records->next ) { # map the oai metadata to mylibrary attributes $fkey = $record->header->identifier; $metadata = $record->metadata; $name = $metadata->title; @creators = $metadata->creator; $note = $metadata->description; $publisher = $metadata->publisher; next if ( ! $publisher ); $location = $metadata->identifier; next if ( ! $location ); $date = $metadata->date; $source = $metadata->source; @subjects = $metadata->subject; # create and commit a mylibrary resource $resource = mylibrary::resource->new; $resource->fkey( $fkey ); $resource->name( $name ); $creator = ‘’; foreach ( @creators ) { $creator .= “$_|” } $resource->creator( $creator ); $resource->note( $note ); $resource->publisher( $publisher ); $resource->source( $source ); $resource->date( $date ); $subject = ‘’; foreach ( @subjects ) { $subject .= “$_|” } $resource->subject( $subject ); $resource->related_terms( new => [ $articleid ]); $resource->add_location( location => $location, location_type => $location_type_id ); $resource->commit; } } information technology and libraries | september # done exit; appendix b # index mylibrary data with kinosearch # require use kinosearch::invindexer; use kinosearch::analysis::polyanalyzer; use mylibrary::core; # define use constant index => ‘../etc/index’; # location of the index mylibrary::config->instance( ‘articles’ ); # mylibrary instance to use # initialize the index $analyzer = kinosearch::analysis::polyanalyzer->new( language => ‘en’ ); $invindexer = kinosearch::invindexer->new( invindex => index, create => , analyzer => $analyzer ); # define the index’s fields $invindexer->spec_field( name => ‘id’ ); $invindexer->spec_field( name => ‘title’ ); $invindexer->spec_field( name => ‘description’ ); $invindexer->spec_field( name => ‘source’ ); $invindexer->spec_field( name => ‘publisher’ ); $invindexer->spec_field( name => ‘subject’ ); $invindexer->spec_field( name => ‘creator’ ); # get and process each resource foreach ( mylibrary::resource->get_ids ) { # create, fill, and commit a document with content my $resource = mylibrary::resource->new( id => $_ ); my $doc = $invindexer->new_doc; $doc->set_value ( id => $resource->id ); mylibrary: a digital library framework and toolkit | morgan $doc->set_value ( title => $resource->name ) unless ( ! $resource->name ); $doc->set_value ( source => $resource->source ) unless ( ! $resource->source ); $doc->set_value ( publisher => $resource->publisher ) unless ( ! $resource->publisher ); $doc->set_value ( subject => $resource->subject ) unless ( ! $resource->subject ); $doc->set_value ( creator => $resource->creator ) unless ( ! $resource->creator ); $doc->set_value ( description => $resource->note ) unless ( ! $resource->note ); $invindexer->add_doc( $doc ); } # optimize and done $invindexer->finish( optimize => ); exit; appendix c # search a kinosearch index and display content from mylibrary # require use kinosearch::searcher; use kinosearch::analysis::polyanalyzer; use mylibrary::core; # define use constant index => ‘../etc/index’; # location of the index mylibrary::config->instance( ‘articles’ ); # mylibrary instance to use # get the query my $query = shift; if ( ! $query ) { print “enter a query. “; chop ( $query = )} # open the index $analyzer = kinosearch::analysis::polyanalyzer->new( language => ‘en’ ); $searcher = kinosearch::searcher->new( invindex => index, analyzer => $analyzer ); # search $hits = $searcher->search( qq( $query )); # get the number of hits and display $total_hits = $hits->total_hits; information technology and libraries | september print “your query ($query) found $total_hits record(s).\n\n”; # process each search result while ( $hit = $hits->fetch_hit_hashref ) { # get the mylibrary resource $resource = mylibrary::resource->new( id => $hit->{ ‘id’ }); # extract dublin core elements and display print “ id = “ . $resource->id . “\n”; print “ name = “ . $resource->name . “\n”; print “ date = “ . $resource->date . “\n”; print “ note = “ . $resource->note . “\n”; print “ creators = “; foreach ( split /\|/, $resource->creator ) { print “$_; “ } print “\n”; # get related terms and display @resource_terms = $resource->related_terms(); print “ term(s) = “; foreach (@resource_terms) { $term = mylibrary::term->new(id => $_); print $term->term_name, “ ($_)”, ‘; ‘; } print “\n”; # get locations (urls) and display @locations = $resource->resource_locations(); print “ location(s) = “; foreach (@locations) { print $_->location, “; “ } print “\n\n”; } # done exit; information technology and libraries | june metadata to support next-generation library resource discovery: lessons from the extensible catalog, phase jennifer bowen the extensible catalog (xc) project at the university of rochester will design and develop a set of open-source applications to provide libraries with an alternative way to reveal their collections to library users. the goals and functional requirements developed for xc reveal gener- alizable needs for metadata to support a next-generation discovery system. the strategies that the xc project team and xc partner institutions will use to address these issues can contribute to an agenda for attention and action within the library community to ensure that library metadata will continue to support online resource discovery in the future. library metadata, whether in the form of marc catalog records or in a variety of newer metadata schemas, has served its purpose for library users by facilitating their discovery of library resources within online library catalogs (opacs), digital libraries, and insti- tutional repositories. however, libraries now face the chal- lenge of making this wealth of legacy catalog data function adequately within next-generation web discovery environ- ments. approaching this challenge will require: n an understanding of the metadata itself and a com- mitment to deriving as much value from it as pos- sible; n a vision for the capabilities of future technology; n an understanding of the needs of current (and, where possible, future) library users; and n a commitment to ensuring that lessons learned in this area inform the development of both future library systems and future metadata standards. the university of rochester ’s extensible catalog (xc) project will bring these various perspectives together to design and develop a set of open-source, collaboratively built next-generation discovery tools for libraries. the xc project team seeks to make the best possible use of legacy library metadata, while also informing the future development of discovery metadata for librar- ies. during phase of the xc project ( – ), the xc project team created a plan for developing xc and defined the goals and initial functional requirements for the system. this paper outlines the major metadata- related issues that the xc project team and xc partner institutions will need to address to build the xc system during phase . it also describes how the xc team and xc partners will address these issues, and concludes by presenting a number of issues for the broader library community to consider. while this paper focuses on the work of a single library project, the goals and functional requirements developed for the xc project reveal many generaliz- able needs for metadata to support a next-generation discovery system. the metadata-related goals of the xc project—to facilitate the use of marc metadata outside an integrated library system (ils), to combine marc metadata with metadata from other sources in a single discovery environment, and to facilitate new functional- ity (e.g., faceted browsing, user tagging)—are very simi- lar to the goals of other library projects and commercial vendor discovery software. the issues described in this paper thus transcend their connection to the xc project and can be considered general needs for library discov- ery metadata in the near future. in addition to informing the library community about the xc project and encouraging comment on that work, the author hopes that identifying and describing meta- data issues that are important for xc—and that are likely to be important for other projects as well—will encourage the library community to set these issues as high priorities for attention and action within the next few years. n the extensible catalog project the university of rochester’s vision for the extensible catalog (xc) is to design and develop a set of open-source applications that provide libraries with an alternative way to reveal their collections to library users. xc will provide easy access to all resources (both digital and physical col- lections) and will enable library content to be revealed through other web applications that libraries may already be using. xc will be released as open-source software, so it will be available for free download, and libraries will be able to adopt, customize, and extend the software to meet their local needs. the xc project is a collaborative effort between partner institutions that will serve a variety of roles in its development. phase of the xc project, funded by the andrew w. mellon foundation and carried out by the university of rochester river campus libraries between april and june , resulted in the creation of a project plan for the development of xc. during xc phase , the xc project team recruited a number of other institutions that will serve as xc partners and who have agreed to contribute resources toward building and implementing xc during phase . xc phase (october through jennifer bowen (jbowen@library.rochester.edu) is director of metadata management at the university of rochester river campus libraries, new york, and is co-principal investigator for the extensible catalog project. metadata to support next-generation library resource discovery | bowen june ) is supported through additional funding from the andrew w. mellon foundation, the university of rochester, and xc partners. during phase , the xc project team, assisted by xc partners, will deploy the xc software and make it available as open-source software. through its various components, the xc system will provide a platform for local development and experimen- tation that will ultimately allow libraries to manage and reveal their metadata through a variety of web applica- tions such as web sites, institutional repositories, and con- tent management systems. a library may choose to create its own customized local interface to xc, or use xc’s native user interface “as is.” the native xc interface will include web . functionality, such as tagging and faceted browsing of search results that will be informed by frbr (functional requirements for bibliographic records) and frad (functional requirements for authority data) conceptual models. the xc software will handle multiple metadata schemas, such as marc and dublin core, and will be able to serve as a repository for both existing and future library metadata. in addition, xc will facilitate the creation and incorporation of user-created metadata, enabling such metadata to be enhanced, augmented, and redistributed in a variety of ways. the xc project team has designed a modular archi- tecture for xc, as shown in the simplified schematic in figure . xc will bring together metadata from a variety of sources (integrated library systems, digital reposito- ries, etc.), apply services to that metadata, and display it in a usable way in the web environments where users expect to find it. xc’s architecture will allow institutions that implement the software to take advantage of innova- tive models for shared metadata services, which will be described in this paper. n xc phase activities during the now-completed xc phase , the xc project team focused on six areas of activity: . survey and understand existing research on user practices. . gauge library demand for the xc system. . anticipate and prepare for the metadata requirements of the new system. . learn about and build on related projects. . experiment with and incorporate useful, freely available code. . build a community of interest. the xc project team carried out a variety of research activities to inform the overall goals and high-level functional requirements for xc. this research included a literature search and ongoing monitoring of discussion lists and blogs, to allow the team to keep up with the most current discussions taking place about next-generation library discovery systems and related technologies and projects. the xc team also consulted regularly with prospective partners and other knowledgeable colleagues who are engaged in defining the concept of a next-gener- ation library discovery system. in order to gauge library demand for the xc system, the team also conducted a survey of interested institutions. this paper reports the results of the third area of activity during xc phase —anticipating and preparing for the metadata requirements of the new system—and looks ahead to plans to develop the xc software during phase . n xc goals and metadata functional requirements the goals of the xc project have significant implications for the metadata functionality of the system, with each goal suggesting specific high-level functional require- ments for how the system can achieve that particular goal. the five goals are: n goal : provide access to all library resources, digital and non-digital. n goal : bring metadata about library resources into a more open web environment. n goal : provide an interface with new web func- tionality such as web . features and faceted browsing. n goal : conduct user research to inform system development. n goal : publish the xc code as open-source software. figure . xc system diagram information technology and libraries | june an overview of each xc goal and its related high-level metadata requirements appears below. each requirement is then discussed in more detail, with a plan for how the xc project team will address that requirement when developing the xc software. n goal : provide access to all library resources, digital and non-digital working alongside a library’s current integrated library system (ils) and its other web applications, xc will strive to bring together access to all library resources, thus eliminating the data silos that are now likely to exist between a library’s opac and its various digital reposi- tories and commercial databases. this goal suggests two fairly obvious metadata requirements (requirements and ). requirement —the system must be capable of acquiring and managing metadata from multiple sources: ilss, digital repositories, licensed databases, etc. a typical library currently has metadata pertaining to its collections residing in a variety of separate online systems: marc data in an ils, metadata in various sche- mas in digital collections and repositories, citation data in commercial databases, and other content on library web sites. a library that implements xc may want to populate the system with metadata from several online environments to simplify access to all types of resources. to achieve goal , xc must be capable of acquiring and managing metadata from all of these sources. each online environment and type of metadata present their own challenges. repurposing marc data repurposing marc metadata from an existing ils will be one of the biggest metadata tasks for a next-generation discovery system such as xc. in planning xc, we have assumed that most libraries will keep their current ils for the next few years or perhaps migrate to a newer commer- cial or open-source ils. in either case, most libraries will likely continue to rely on an ils’s staff functionality to handle materials acquisition, cataloging, circulation, etc. for the short term. relying upon an ils as a processing environment does not, however, mean that a library must use the opac portion of that ils as its means of resource discovery for users. xc will provide other options for resource retrieval by using web services to interact with the ils in the background. to repurpose ils metadata and enable it to be used in various web discovery envi- ronments, xc will harvest a copy of marc metadata records from an institution’s ils using the open archives initiative protocol for metadata harvesting (oai-pmh). using web services and standard protocols such as oai- pmh offers not only a short-term solution for reusing metadata from an ils, but can also be used in both the short- and long-term to harvest metadata from any sys- tem that is oai-pmh harvestable, as will be discussed further below. while harvesting metadata from existing systems into xc creates duplication of metadata between an ils and xc, this actually has significant benefits. xc will handle metadata updates through automated harvesting services that minimize additional work for library staff, other than for setting up and managing the automated services themselves. the internal xc metadata cache can be easily regenerated from the original repositories and services when necessary, such as to enable future changes to the internal xc metadata schema. the xc system architecture also makes use of internal metadata duplication among xc’s components, which allows these components to communicate with each other using oai- pmh. this built-in metadata redundancy will also enable xc to communicate with external services using this standard protocol. it is important to distinguish the deliberate metadata redundancies built into the xc architecture from the type of metadata redundancies that have been singled out for elimination in the library of congress working group on the future of bibliographic control draft report (recommendation . ) and previously in the university of california (uc) libraries bibliographic services task force’s final report. these other “negative” redundan- cies result from difficulties in sharing metadata among different environments and cause significant additional staff expense for libraries to enrich or recreate metadata locally. xc’s architecture actually solves many of these problems by facilitating the sharing of enriched metadata among xc users. xc can also adapt as the library com- munity begins to address the types of costly metadata redundancies mentioned in the above reports, such as between the oclc worldcat database and copies of that marc data contained within a library’s ils, because xc will be capable of harvesting metadata from any source that uses a standard api. metadata from digital repositories and other free sources xc will harvest metadata from various digital collections and repositories, using oai-pmh, and will maintain a copy of the harvested metadata within the xc metadata cache, as shown in figure . the metadata services hub architecture provides flexibility and possible economy for xc users by offering the option for multiple xc insti- tutions to share a single metadata hub, thus allowing participating institutions to take full advantage of the hub’s capabilities to aggregate and augment metadata from multiple sources. while the procedure for harvest- metadata to support next-generation library resource discovery | bowen ing metadata from an external repository is not techno- logically difficult in itself, managing the flow of metadata coming from multiple sources and aggregating that meta- data for use in xc will require the development of sophis- ticated software. to address this, the xc project team is partnering with established experts in bibliographic metadata aggregation to develop the metadata services portion of the xc architecture. the team from cornell university that has developed the software behind the national science digital library’s metadata management system (nsdl/mms) is advising the xc team in the development of the xc metadata services hub, which will be built on top of the basic nsdl/mms software. the xc metadata services hub will coordinate meta- data services into a reusable task grouping that can be started on demand or scheduled to run regularly. this xc component will harvest xml metadata and com- bine metadata records that refer to equivalent resources (based on uniform resource identifier [uri], if available, or other unique identifier) into what the cornell team describes as a “mudball.” each mudball will contain the original metadata, the sources for the metadata, and the references to any services used to combine metadata into the mudball. the mudball may also contain metadata that is the result of further automated processing or services to improve quality or to explicitly identify relationships between resources. hub services could potentially record the source of each individual metadata statement within each mudball, which would then allow a metadata record to be redelivered in its original or in an enriched form when requested. by allowing for the capture of prov- enance data for each data element, the hub could poten- tially provide much more granular information about the origin of metadata—and much more flexibility for recombining metadata—than is possible in most marc- based environments. after using the redeployed nsdl/mms software as the foundation for the xc metadata hub, the xc project team will develop additional hub services to support xc’s functional requirements. xc-specific hub services will accommodate incoming marc data (including marc holdings data for non-digital resources); basic authority control; mappings from marc , marcxml, and dublin core to an internal xc schema defined within the xc application profile (described below); and other ser- vices to facilitate the functionality of the xc user environ- ments (see discussion of requirement , below). finally, the xc hub services will make the metadata available for harvesting from the hub by the xc client integration applications. metadata for licensed content for a next-generation discovery system such as xc to provide access to all library resources, it will need to pro- vide access to licensed content, such as citation data and full-text databases. metasearch technology provides one option for incorporating access to licensed content into xc. unfortunately, various difficulties with metasearch technology and usability issues with some metasearch products make metasearch technology a less-than-ideal solution. an alternative approach would bring meta- data from licensed content directly into a system such as xc. the metadata services hub architecture for xc is capable of handling the ingest and processing of meta- data supplied by commercial content providers by add- ing additional services to handle the necessary schema transformations and to control access to the licensed content. the more difficult issue with licensed content may be to obtain the cooperation of commercial vendors to ingest their metadata into xc. pursuing individual agreements with vendors to negotiate rights to ingest their metadata is beyond the original scope of xc’s phase project. however, the xc team will continue to monitor ongoing developments in this area, especially the work of the ethicshare project, which uses a system architecture very similar to that of xc. it remains our goal to build a system that will facilitate the inclusion of licensed content within xc in situations where commercial providers have made it available to xc users. requirement summary when considering needed functionality for a next-gen- eration discovery system, the ability to ingest and man- age metadata from a variety of sources is of paramount importance. unlike a current ils, where we often think of metadata as mostly static unless it is supplemented by new, updated, and deleted records, we should instead envision the metadata in a next-generation system as being in constant motion, moving from one environment to another and being harvested and transformed on a scheduled basis. the metadata services hub architecture of the xc system will accommodate and facilitate such constant movement of metadata. requirement —the system must handle multiple metadata schemas. an extension of requirement will be the necessity for a next-generation system such as xc to handle metadata from multiple schemas, as the system harvests those sche- mas from various sources. library metadata priorities as a part of the xc survey of libraries described earlier in this paper, the xt team queried respondents about what metadata schemas they currently use or plan to use in the near future. many responding libraries indicated that they expect to increase their use of non–marc metadata within the next three years, although no library indicated the intention to completely move away from information technology and libraries | june marc within that time period. nevertheless, the idea of a “marc exit strategy” has been discussed in various circles. the architecture of xc will enable libraries to move beyond the constraints of a marc-based system without abandoning their ils, and will provide an oppor- tunity for libraries to stage their “marc exit strategy” in a way that suits their purposes. libraries also indicated that they plan to move away from homegrown schemas toward accepted standards such as mets, mods, mads, premis, ead, vra core, and dublin core. several responding libraries plan to move toward a wider variety of metadata schemas in the near future, and will focus on using xml- based schemas to facilitate interoperability and metadata harvesting. to address the needs of these libraries in the future, xc’s metadata services will contain a variety of transformation services to handle a variety of schemas. taking into account the metadata schemas mentioned the most often among survey respondents, the software developed during phase of the xc project will sup- port harvested metadata in marc , marcxml, and dublin core (including qualified dublin core). metadata crosswalks and mapping one respondent to the xc survey offered the prediction that “reuse of existing metadata and transformation of metadata from one format to another will become commonplace and routine.” xc’s internal metadata transformations must be designed with this in mind, to facilitate making these activities “commonplace and routine.” fortunately, many maps and crosswalks already exist that potentially can be incorporated into a next-generation system such as xc. the metadata ser- vices hub architecture for xc can function as a standard framework for applying a variety of existing crosswalks within a single, shared environment. following “best practices” for crosswalking metadata, such as those developed by the digital library federation (dlf), will be extremely important in this environment. as the dlf guidelines describe, metadata schema transforma- tion is not as straightforward as it might first appear to be. while the dlf guidelines advise always cross- walking from a more robust schema to a simpler one, sometimes in a series of steps, such mapping will often result in “dumbing down” of metadata, or loss of granu- larity. this is a particularly important concern for the xc project because a large percentage of the metadata handled by xc will be rich legacy marc metadata, and we hope to maintain as much of that richness as possible within the xc system. in addition to simply mapping one data element in a schema to its closest equivalent in another, it is essential to ensure that the underlying metadata models of the two schemas being crosswalked are compatible. the authors of the framework for a bibliographic future draft document define multiple layers of such models that need to be considered, and offer a general high- level comparison between the frbr data model and the dcmi (dublin core metadata initiative) abstract model (dcam). more detailed comparisons of models are also taking place as a part of the development of the new metadata content standard, resource description and access (rda). the developers of rda have issued documents offering a detailed mapping of rda elements to rda’s underlying model (frbr) and analyzing the relationship between rda elements, the dcmi abstract model, and the metadata framework. as a result of a meeting held april –may , , a joint dcmi/rda task group is now undertaking the collaborative work necessary to carry out the following tasks: n develop an rda element vocabulary. n develop an rda/dublin core application profile based on frbr and frad. n disclose rda value vocabularies using rdf/ rdfs/skos. these efforts hold much potential to provide a more rigorous way to communicate about metadata across multiple communities and to increase the compatibility of different metadata schemas and their underlying models. such compatibility will be essential to enabling the func- tionality of future discovery systems such as xc. an xc metadata application profile the xc project team will define a metadata application profile for xc as a way to document decisions made about data elements, content standards, and crosswalking used within the system. the use of an application profile can facilitate metadata migration, harvesting, and other auto- mated processes, and presents an approach to metadata that is more flexible and responsive to local needs than simply adopting someone else’s metadata guidelines. application profiles facilitate the use of multiple schemas because elements can be selected for inclusion from more than one existing schema, or additional elements can be created and defined locally. because the xc system will incorporate harvested metadata from a variety of sources, the use of an application profile will be essential to sup- port xc’s complex system requirements. the dcmi community has published guidelines for creating a dublin core application profile (dcap), which is defined more specifically as: [a] form for documenting which terms a given applica- tion uses in its metadata, with what extensions or adap- tations, and specifying how those terms relate both to formal standards such as dublin core as well as to less formally defined element sets and vocabularies. metadata to support next-generation library resource discovery | bowen the announcement of plans to develop an rda/ dublin core application profile illustrates the impor- tant role that application profiles are beginning to take to facilitate the interoperability of metadata schemas. the planned rda/dc application profile will “trans- late” rda into a standard structure that will allow it to be related more easily to other metadata element sets. unfortunately, the rda/dc application profile will likely not be completed in time for it to be incorporated into the first release of the xc software in mid- . nevertheless, we intend to use the existing definitions of rda elements to inform the development of the xc application profile. this will allow us to anticipate any future incompatibilities between the rda/dc and the xc application profiles, and ensure that xc will be well- positioned to take advantage of rda-based metadata when rda is implemented. this process may have the reciprocal benefit of also informing the developers of rda of any rda elements that may be difficult to imple- ment within a next-generation system such as xc. the potential value of rda to the xc project—in terms of providing a consistent approach to bibliographic and authority metadata and facilitating frbr-related user functionality—is very significant. it is hoped that at some point xc can become an early adopter of rda and provide a mechanism through which libraries can move their legacy marc metadata into a system that is compatible with an emerging international metadata standard. n goal : bring metadata about library resources into a more open web environment xc will reveal library metadata not only through its own separate interface (either the out-of-the-box xc interface or an interface designed by the local library), but will also allow library metadata to be revealed through other web applications. the latter approach will bring library resources directly to web locations that library users are already visiting, rather than attempting to entice users to visit an additional library-specific web location. making library metadata work effectively in the broader web environment (outside the well-defined boundar- ies of an ils or repository) will require the following requirements and : requirement —metadata must conform to the standards of the new web environments as well as to that of the system from which it originated. achieving requirement will require library metadata in future systems to perform a dual function: to conform to both existing library standards as well as to web standards and conventions. one way to achieve this is to ensure that the two types of standards themselves are compatible. coyle and hillmann have argued persua- sively for changes in the direction of rda development to allow metadata created using rda to function in the broader web environment. these changes include the need to follow a clearly refined, high-level metadata model, to create data elements that can be manipulated by machines, and to move toward the use of uris instead of textual identifiers. after the announcement of the outcomes of the rda/dc data modeling meet- ing, the two authors are considerably more optimistic about rda functioning as a standard within the broader web environment. this discourse concerning rda shows but a piece of the process through which long-es- tablished library metadata standards need to be reexam- ined to make library metadata understandable to both humans and machines on the web. moving away from aacr toward rda, and ultimately toward incorporat- ing standard web conventions into library metadata, can be a difficult process for those involved in creating and maintaining library standards. nevertheless, transform- ing library metadata standards in this way is essential to fulfill the requirements necessary for next-generation library discovery systems. requirement —metadata must function effectively within the new web environments as well as within the system from which it originated. not only must metadata for a next-generation system follow the conventions and standards used in the broader web, but the data also needs to be able to func- tion effectively in a broader web environment. this is a slightly different proposition from requirement , and will necessitate testing the metadata standards them- selves to ensure that they enable library metadata to function effectively. the xc project will provide direct experience with using library metadata in two types of web environ- ments: content management systems and learning man- agement systems. library metadata in a content management system as shown in the xc architecture diagram in figure , the xc project team will build one of the primary user environments for xc on top of the open-source content management system, drupal. the xc drupal module will allow us to respond to many of the needs expressed by libraries in their responses to the xc survey by supplying: n a web application server with a back-end database; information technology and libraries | june n a user interface with web . features; n library-controlled web pages that will treat library metadata as a native data type; n a metadata interface for enhancing or correcting metadata in the system; and n an administrative interface. the xc team will bring library metadata into the drupal content management system (cms) as a native content type within that environment, creating a drupal “node” for each metadata record. this will allow xc to take advantage of many native features of the drupal cms, such as a taxonomy system. building xc inter- faces on top of the drupal cms will also give us an opportunity to collaborate with partner libraries that are already active participants in the drupal user com- munity. xc’s architecture will allow the possibility of develop- ing additional user environments on top of other content management systems. bringing library metadata into these new environments will provide many new oppor- tunities for libraries to manipulate their metadata and present it to users without being constrained by the limi- tations of the current generation of library systems. such opportunities will then inform the future requirements for library metadata in such environments. library metadata in a learning management system figure illustrates two examples of xc user envi- ronments through learning management systems: xc interfaces to both the blackboard learning system and sakai. much exciting work is being done at other institutions to bring library content into these web applications. xc will build on projects such as these to reveal library metadata for non-licensed library resources from an ils through learning management systems. specifically, we plan to develop the capabil- ity for libraries to make the display of library metadata context-sensitive within the learning management sys- tem. for example, searching or browsing on a page for a particular academic course could be configured to reflect the subject area of the course (e.g., chemistry) and automatically present library resources related to that subject. this capability will build upon the experiences gained by the university of rochester through its work to develop its “course resources” system. such xc functionality will be integrated directly into the learn- ing management system, rather than simply providing a link out to a separate library system. again, we hope that our efforts to bring library metadata into these new environments will encourage libraries to engage in further work to integrate library resources into broader web environments and inform future requirements for library metadata in these envi- ronments. n goal : provide an interface with new web functionality such as web . features and faceted browsing new functionality for users will require that metadata fulfill more sophisticated functions in a next-generation system than it may have done in an ils or repository, in order to provide more intuitive searching and navigation. the system will also need to capture and incorporate metadata generated through tagging, user-contributed reviews, etc. such new functionality creates the need for requirements and . requirement —metadata must support functionality to facilitate intuitive searching and navigation, such as faceted browsing and frbr- informed results groupings. enabling faceting and clustering much research has already been done regarding the design of faceted search interfaces in general. when con- sidered along with user research conducted at other insti- tutions and to be conducted during the development of xc, this data provides a strong foundation for the design of a faceted browse environment. the xc project team has already gained firsthand experience with developing faceted browsing through the development of the “c ” prototype interface during phase of the xc project. to enable faceting within xc, we will also pay particular attention to what others have discovered through design- ing faceted interfaces on top of legacy marc meta- data. specific lessons learned from those involved with north carolina state university’s endeca-based catalog, vanderbilt university’s primo implementation, and plymouth state university’s scriblio system provide valuable guidance for the xc project team as we design facets for the xc system. ideally, a mechanism should be developed to enable these discoveries to feed back into the development of metadata and encoding standards, so that changes to existing standards can be considered to facilitate faceting in the future. several new system implementations have used library of congress subject headings (lcsh) and lc subdivisions from marc records as the basis for deriving facets. the xc “c ” prototype interface provides facets for topic, genre, and region that are based simply upon one or more marc xx tags. north carolina state university’s endeca-based system has enabled facets for topic, genre, region, and era using lcsh subdivisions as well, but this has necessitated a “massive cleanup” of subdivisions, as described by charley pennell. oclc’s fast (faceted application of subject terminology) project may provide another option for enabling such facets. a library could populate its marc data with fast headings, based metadata to support next-generation library resource discovery | bowen upon the existing lcsh in the records, and then use the fast headings as the basis for generating facets. it remains to be seen whether fast will offer significant benefit over lcsh itself when it comes to faceting, however, since fast headings are generated directly from lcsh. while marc metadata has some known difficul- ties where faceting and clustering are concerned (such as those involving lcsh), the xc system will encounter additional difficulties when implementing these tech- nologies with less robust metadata schemas such as simple dublin core, and especially across metadata from a variety of schemas. the development of web services to augment batches of metadata records in an automated manner holds some promise for improving the creation of facets from other metadata schemas. within the xc system, such services could be added to the metadata services hub and run against ingested metadata. while designing extensive services of this type is beyond the scope of the next phase of xc software development, we will encourage others to develop such services for xc. another (but much less desirable) approach to aug- menting metadata is for a metadata specialist to manually edit one record or group of records. the xc cataloging interface, built within the drupal cms, will allow record- by-record editing of metadata when necessary. while we see this editing interface as essential functionality for xc, we anticipate that libraries will want to use this feature sparingly. in many cases it will be preferable to correct or augment metadata within its original repository (e.g., the institution’s ils) and then re-harvest the corrected meta- data, rather than correcting it manually within xc itself. because of the expense of manual metadata augmentation and correction, libraries will be well-advised to rely upon insights gained through user research to assess the value of this type of work. for example, a library might decide to edit individual metadata records only when the correction or augmentation will support specific system functionality that is of high priority for the institution’s users. implementing frbr results groupings to incorporate logical groupings of search results based upon the frbr and frad data models over sets of diverse metadata within xc, we will encounter similar difficulties that we face with faceting and clustering. various analyses of the marc formats have dealt extensively with the relationship between frbr and marc , and others have written specifically about methodology for frbrizing a marc-based catalog. in addition, various tools and web services are available that can potentially facilitate this process. even with this extensive body of work to draw upon, however, the suc- cess of our implementation of frbr-based functionality will depend upon both the quality and completeness of the system’s metadata. metadata in xc that originated as dublin core records may need significant augmenta- tion to be incorporated effectively into frbrized results displays. to maximize the ability of the system to support frbr/frad results groupings, we may need to supple- ment automated grouping of resources with a combina- tion of additional services for the metadata services hub, and with cataloger-generated metadata correction and augmentation, as described above. the xc team will use the results of user research carried out during the next phase of the xc project to inform our decision-making regarding what frbr-informed results grouping users find helpful, and then assess what specific metadata aug- mentation services are needed for xc. providing frbr-informed groupings of related records in search results will be easier when the underly- ing metadata incorporates principles of authority control. of course, the vast majority of the non-marc metadata that will be ingested into xc will not be under author- ity control. again, this situation suggests the need for additional services or functionality to improve existing metadata within the xc metadata hub, the xc cataloging interface, or both. as an experiment in developing ser- vices to facilitate authority control, the xc project team carried out a pilot project in partnership with a group of software engineering students from the rochester institute of technology (rit) during phase of xc. the rit students designed a basic name access control tool that can be used across disparate metadata schemas in an environment such as xc. the tool can ingest marc authority and bibliographic records as well as dublin core records, provide automated matching, and facili- tate a cataloger’s handling of problem reports. the xc project team will implement the automated portion of the tool as a web service within the xc hub, and the “cataloger facilitation” portion of the tool within the xc cataloging user interface. institutions that use xc can then incorporate additional tools to facilitate authority control into xc as they are needed and developed. in addition to providing a test case for developing xc metadata services, the rit pilot project proved valuable by providing an opportunity for student software devel- opers and catalogers to discuss the functional require- ments of a cataloging tool. not only did the experience enable the developers to understand the needs of the system’s intended users, but it also presented an opportu- nity for the engineering students to demonstrate techno- logical possibilities that the catalogers—who work almost exclusively with legacy ils technology—may not have envisioned before participating in the project. requirement —the system must manage user- generated metadata resulting from user tagging, submission of reviews, etc. because users now expect web-based tools to offer web . functionalities, the xc project has as one of its basic information technology and libraries | june goals to incorporate these functionalities into xc’s user environments. the results of the xc survey rank tools to support the finding, gathering, use, and reuse of scholarly content (e.g., rss feeds, blogs, tagging, user reviews) eighth out of a list of twenty new desirable opac fea- tures. we expect to learn much more about the useful- ness of web . technology within a next-generation system through the user research that we will carry out during phase of the xc project. the xc system will capture metadata generated by users from any one of the system’s user environ- ments (e.g., drupal-based interface, learning manage- ment system integration) and harvest it back into the system’s metadata services hub for processing. the xc application profile will incorporate user-generated metadata, mapped into its own carefully defined meta- data elements. this will allow us to capture and manage this metadata as discrete content, without inadvertently mixing it with other metadata created by library staff or ingested from other sources. n goal : conduct user research to inform system development user research will be essential to informing the design and functionality of the xc software. to align xc’s functional requirements as closely as possible with user needs, the xc project team will practice a user-centered design methodology that takes an iterative approach to defining the system’s functional requirements. since we will engage concurrently in the processes of user research and software design, we will not fully determine the system requirements for xc until a significant amount of user research has been done. a complete picture of the demands upon metadata within xc will thus emerge as we gain information from our user research. n goal : publish the xc code as open-source software central to the vision of the xc project is sharing the xc software freely throughout the library community and beyond. our hope is that others will use all or part of the xc software, modify it, and improve it to meet their own needs. new requirements for the metadata within xc are likely to arise as this process takes place. other future changes to the xc software will also be needed to ensure the software’s continued compatibility with vari- ous metadata standards and schemas. these changes will all affect the system requirements for xc over time. addressing goals and while goals through for the xc project result in specific high-level functional requirements for the sys- tem’s discovery metadata that can be addressed and dis- cussed as xc is being developed, goals and present general challenges that must be addressed in the future. goal is likely to fuel the need to update the xc software over time as the needs of users change. goal provides a challenge to managing that updating process in a col- laborative environment. these two goals suggest an additional general requirement for the system’s metadata requirement : requirement —the system’s metadata must be extensible to facilitate future enhancements and updates. enabling future user needs developing xc using a user-centered design process in which user research and software design occur simulta- neously will enable us to design and build a system that is as responsive as possible to the needs of users that are seeking library resources. however, user needs will change during the life of the xc software. these needs must be assessed and addressed, and then weighed against the desires of individual institutions that use xc and who request specific system enhancements. to carry forward the xc project’s commitment to serving users, we will develop a governance model for the xc community that brings the needs of future users into the decision-making process by providing a method for continuing to determine and capture user needs. in addition, we will consciously cultivate a commitment to user research among members of the xc community. because the xc software will be released as open source, we can also encourage xc partners to develop whatever additional functionality they need for their own insti- tutions and make these enhancements available to the entire community of xc users. this approach is very different from the enhancement process in place for most commercial systems, and xc partner institutions may need to adjust to this approach. enabling future metadata standards as current metadata standards are revised and new standards and schemas are created, xc must be able to accommodate these changes. new crosswalks will allow new metadata schemas to be mapped to the xc internal schema in the future. the xc application profile can be updated with the addition of new data elements as needed. the drupal-based xc user environment will also allow institutions that use xc to create new internal data types to incorporate additional types of metadata. as the development of the semantic web moves forward and enables smart linking between existing authority files and vocabularies, xc’s architecture can make use of the resulting web services, either by incorporating them metadata to support next-generation library resource discovery | bowen through the xc metadata services hub or through the native xc user interface as part of a user search query. n further considerations the above discussion of the goals and requirements for xc has revealed a number of issues related to the devel- opment of next-generation discovery systems that are unfortunately beyond the scope of the next phase of the xc project. we therefore offer them as a possible agenda for future work by the broader library community: . explore the wider usefulness of web-based meta- data services and the need for an automated metadata services coordinator to control these functions. libraries are already comfortable with basic “services” that are performed on metadata by an outside agency: for example, a library may send copies of its marc records to a vendor for authority processing or enrichment with tables of contents or other data elements. the library com- munity should encourage vendors and others to develop these and other metadata enrichment options as automated web services. . study the advantages of using statement-level metadata provenance, as used in the nsdl metadata management system and considered for use within the xc metadata services hub, and explore whether there are ways that marc could move toward allowing more granularity in recording and sharing metadata provenance. . to facilitate access to licensed library resources, encourage the development of more robust metase- arch technology and standards so that technologi- cal limitations do not hinder system performance and search result usability. if this is not successful, libraries and content providers must work together to enable metadata for licensed resources to be revealed within open discovery environments such as xc and ethicshare. this second scenario will enable libraries to directly address usability issues with the display of licensed content, which may make it a more desirable longer-term solution than attempting to improve metasearch technology. . the administrative bodies of the two groups rep- resented on the dcmi/rda task group (i.e., the dublin core metadata initiative and the rda committee of principals) have a responsibility to take the lead in funding this group’s work to develop and maintain the rda/dc application profile and its related registries and vocabularies. beyond this, however, the broader library com- munity must recognize that this work is essential to ensure that future library metadata standards will function in the broader web environment, and offer additional administrative and financial sup- port for it in the coming years. . to ensure that library standards work effectively outside of traditional library systems, catalog- ers and metadata experts must develop ongoing, collaborative working relationships with system developers. such collaboration will necessitate educating each group of experts about the domain of the other. . libraries should experiment with using metadata in new environments and use the lessons learned from this activity to inform the metadata standards development process. while current library auto- mation environments by and large do not provide opportunities for this, the extensible catalog will provide a flexible platform where experimenta- tion can take place. xc will make experimenta- tion as risk-free as possible by ensuring that the original metadata brought into the system can be reharvested in its original form, thus minimizing concerns about possible data corruption. xc will also minimize the investment needed for a library to engage in this experimentation because it will be released as open-source software. . to facilitate new functionality for next-generation library discovery environments, libraries must share their new expertise in this area with each other. for example, library professional organiza- tions (such as ala and its associations) should form discussion groups and committees devoted to sharing lessons learned from the implementa- tion of faceted interfaces and web . technologies, such as tagging and folksonomies. such groups should develop a “best practices” document out- lining a preferred way to define facets from marc data that can be used by any library implement- ing faceting on top of its legacy metadata. . the library community should discuss and encour- age mechanisms for pooling and sharing user- generated metadata among libraries and other interested institutions. n conclusions to present library resources via the web in a manner that users now expect, library metadata must function in ways that have never been required of it before. making library metadata function effectively within the broader web environment will require that libraries take advantage of the combined knowledge of experts in the areas of cata- loging/metadata and system development who share a information technology and libraries | june common vision for serving library users. the challenges to making legacy library metadata and newer metadata for digital resources interact effectively in the broader web environment are significant, and work must begin now to ensure that we can preserve the investment that libraries have made in their legacy metadata. while the recommendations within this report are the result of planning to develop one particular library discovery system—the extensible catalog (xc)—these lessons can inform the development of other systems as well. the actual development of xc will continue to add to our knowledge in this area. while it may be tempting to wait and see what commercial vendors offer as their next generation of commercial discovery products, such a passive approach may jeopardize the future viability of library metadata. projects such as the extensible catalog can serve as a vehicle for moving forward by providing an opportunity for libraries to experiment and to then take informed action to move the library community toward a next generation of resource discovery systems. acknowledgments phase of the extensible catalog project was funded through a grant from the andrew w. mellon foundation. this paper is in partial fulfillment of that grant, originally funded on april , , and concluding on june , . the author acknowledges the contributions of the entire university of rochester extensible catalog project team to the content of this paper, and especially thanks david lindahl, barbara tillett, and konstantin gurevich for reading and offering suggestions on drafts of this paper. references and notes . despite the use of the word “catalog” within the name of the extensible catalog project, this paper will avoid using the word “catalog” in the phrase “next-generation catalog” because this may misleadingly convey the idea of a catalog as solely a single, separate web destination for library users. instead, terms such as “discovery environment” and “discovery system” will be preferred. . the xc blog provides a list of xc partners, describes their roles in xc phase , and provides links to reports that represent the outcomes of xc phase . “xc (extensible catalog): an open- source online system that will unify access to traditional and digital library resources,” www.extensiblecatalog.info (accessed october , ). . ifla study group on the functional requirements for bibliographic records, functional requirements for bibliographic records (munich: k. g. saur, ), www.ifla.org/vii/s /frbr/ frbr.pdf (accessed july , ). . ifla working group on functional requirements and numbering of authority records (franar), “functional requirements for authority data: a conceptual model,” april , , www.ifla.org/vii/d /franar-conceptualmodel- ndreview.pdf (accessed july , ). . library of congress, network development and marc standards office, “marc formats,” april , , www.loc .gov/marc/marcdocz.html (accessed september , ). . “dublin core metadata element set, version . ,” decem- ber , , http://dublincore.org/documents/dces (accessed september , ). . university of rochester river campus libraries, “exten- sible catalog phase ,” (grant proposal submitted to the andrew w. mellon foundation, july , ). . “literature list,” extensible catalog blog, www. extensiblecatalog.info/?page_id= (accessed august , ). . a summary of the results of this survey is available on the xc blog. nancy fried foster et al., “extensible catalog survey report,” july , , www.extensiblecatalog.info/wp-content/ uploads/ / /xc% survey% report.pdf (accessed july , ). . lorcan dempsey has written of the need for a service layer for libraries that would facilitate the “de-coupling” of resource retrieval from back-end processing. lorcan dempsey, “a palindromic ils service layer,” lorcan dempsey’s weblog, january , , http://orweblog.oclc.org/archives/ . html (accessed august , ). . “open archives initiative protocol for metadata harvest- ing v. . ,” www.openarchives.org/oai/openarchivesprotocol. html (accessed august , ). . library of congress, working group on the future of bibliographic control, “report on the future of bibliographic control: draft for public comment,” november , , www .loc.gov/bibliographic-future/news/lcwg-report-draft- - - -final.pdf (accessed december , ). . university of california libraries bibliographic services task force, “rethinking how we provide bibliographic services for the university of california,” final report, , http://libraries. universityofcalifornia.edu/sopag/bstf/final.pdf (accessed august , ). . “[worldcat.org] search for an item in libraries near you,” www.worldcat.org (accessed august , ). . oclc’s plan to create additional apis to worldcat as part of its worldcat grid project is a welcome development that may enable oclc members to harvest metadata directly from worldcat into a system such as xc in the future. see the following blog posting for an early description of oclc’s plans, which have not been formally unveiled by oclc as of this writing: bess sadler, “the librarians and the choco- late factory: oclc developer network day,” solvitur ambu- lando, october , , www.ibiblio.org/bess/?p= (accessed december , ). . “metadata management system,” nsdl registry, sep- tember , , http://metadataregistry.org/wiki/index.php/ metadata_management_system (accessed july , ). . diane hillmann, stuart sutton, and jon phipps, “nsdl metadata improvement and augmentation services,”(grant proposal submitted to the national science foundation, ). . library of congress, network development and marc standards office, “marcxml: marc xml schema,” july , , www.loc.gov/standards/marcxml (accessed september , ). metadata to support next-generation library resource discovery | bowen . andrew k. pace, “category: metasearch,” hectic pace, http://blogs.ala.org/pace.php?cat= (accessed august , ). see in particular the following blog entries: “metameta,” july , ; “more meta,” september , ; “preaching to the publishers,” oct , ; “even more meta,” july , ; and “still here,” august , . . david lindahl, “metasearch in the users’ context,” the serials librarian , no. / ( ): – . . ethicshare, a collaborative project of the university of minnesota, georgetown university, indiana university–bloom- ington, indiana university–purdue university indianapolis, and the university of virginia, is addressing this challenge as part of its plan to develop a sustainable online environment for the practical ethics community. the architecture of the proposed ethicshare system has many similarities to that of xc, but the project focuses specifically upon ingesting citation metadata from a variety of sources, including commercial providers. see cecily marcus, “ethicshare planning phase final report,” july , www.lib.umn.edu/about/ethicshare/university% of% minnesota_ethicshare_final_report.pdf (accessed august , ). . roy tennant used this phrase in “marc exit strategies,” library journal , no. (november , ), www.libraryjour- nal.com/article/ca .html?q=tennant+exit (accessed july , ); karen coyle presented her vision for moving beyond marc to a more flexible, identifier-based record structure that will facilitate a range of library functions in “future consider- ations: the functional library systems record,” library hi tech , no. ( ). . library of congress, network development and marc standards office, “mets: metadata encoding and transmission standard official web site,” august , , www.loc.gov/ standards/mets (accessed september , ). . library of congress, network development and marc standards office, “mods: metadata object description schema,” august , , www.loc.gov/standards/mods (accessed sep- tember , ). . library of congress, network development and marc standards office, “mads: metadata authority description schema,” february , , www.loc.gov/standards/mads (accessed september , ). . “premis: preservation metadata maintenance activity,” july , , www.loc.gov/standards/premis (accessed sep- tember , ). . library of congress, network development and marc standards office, “ead: encoded archival description version official site,” august , , www.loc.gov/ead (accessed september , ). . visual resources association, “vra core: welcome to the vra core . ,” www.vraweb.org/projects/vracore (accessed september , ). . “dublin core metadata element set, version . .” . other xml-compatible schemas, such as mods and mads, will also be supported initially in xc if they are first con- verted into marc xml or qualified dublin core. in the future, we plan to allow these other schemas to be harvested directly into xc. . foster et al., “extensible catalog survey report,” july , , . the original comment was submitted by meg bellinger in yale university’s response to the xc survey. . patricia harpring et al., “metadata standards cross- walks,” in introduction to metadata: pathways to digital informa- tion (getty research institute, n.d.), www.getty.edu/research/ conducting_research/standards/intrometadata/crosswalks. html (accessed august , ); see also carol jean godby, jef- frey a. young, and eric childress, “a repository of metadata crosswalks,” d-lib magazine , no. (december ), www .dlib.org/dlib/december /godby/ godby.html (accessed july , ). . digital library federation, “crosswalkinglogic,” june , , http://webservices.itcs.umich.edu/mediawiki/oaibp/ index.php/crosswalkinglogic (accessed august , ). . karen coyle et al., “framework for a bibliographic future,” may , http://futurelib.pbwiki.com/framework (accessed july , ). . ifla study group on the functional requirements for bibliographic records, functional requirements for bibliographic records. . andy powell et al., “dcmi abstract model,” dublin core metadata initiative, june , , http://dublincore.org/ documents/abstract-model (accessed august , ). . joint steering committee for development of rda, “rda: resource description and access: background,” july , , www.collectionscanada.ca/jsc/rda.html (accessed august , ). . joint steering committee for development of rda, “rda-frbr mapping,” june , , www.collectionscanada .ca/jsc/docs/ rda-frbrmapping.pdf (accessed august , ). . joint steering committee for development of rda, “rda element analysis,” june , , www.collectionscanada.ca/ jsc/docs/ rda-elementanalysis.pdf (accessed august , ). a revised version of the document was issued on december , , at www.collectionscanada.gc.ca/jsc/docs/ rda-element analysisrev.pdf (accessed december , ). . “data model meeting: british library, london april– may ,” www.bl.uk/services/bibliographic/meeting.html (accessed july , ). the task group has outlined its work plan, including deliverables, on its wiki at http://dublincore .org/dcmirdataskgroup (accessed october , ). . emily a hicks, jody perkins, and margaret beecher mau- rer, “application profile development for consortial digital libraries,” library resources and technical services , no. (april ). . makx dekkers, “application profiles, or how to mix and match metadata schemas,” cultivate interactive, january , www.cultivate-int.org/issue /schemas (accessed august , ). . thomas baker et al., “dublin core application profile guidelines,” september , , http://dublincore.org/usage/ documents/profile-guidelines (accessed october , ). . joint steering committee for development of rda, “rda element analysis.” . karen coyle and diane hillmann, “resource descrip- tion and access (rda): cataloging rules for the th century,” d-lib magazine , no. / (jan./feb. ), www.dlib.org/dlib/ january /coyle/ coyle.html (accessed august , ). . karen coyle, “astonishing announcement: rda goes . ,” coyle’s information, may , , http://kcoyle.blogspot .com/ / /astonishing-announcement-rda-goes- .html (accessed august , ). information technology and libraries | june . “drupal.org,” http://drupal.org (accessed august , ). . foster et al., “extensible catalog survey report,” . . “taxonomy: a way to organize your content,” drupal.org, http://drupal.org/handbook/modules/taxonomy (accessed september , ). . “blackboard learning system,” www.blackboard.com/ products/academic_suite/learning_system/index.bb (accessed august , ). . “sakai: collaboration and learning environment for edu- cation,” http://sakaiproject.org (accessed august , ). . for example, the library into blackboard project at california state fullerton has developed a toolkit for faculty that brings openurl resolver functionality into blackboard to create linked citations to resources. see “putting the library into black- board: a toolkit for cal state fullerton faculty,” , www .library.fullerton.edu/librarytoolkit/default.shtml (accessed august , ); and susan tschabrun, “putting the library into blackboard: using the sfx openurl generator to create a toolkit for faculty.” the sakaibrary project at indiana uni- versity and the university of michigan are working to integrate licensed library content into sakai using metasearch technology. see “sakaibrary: integrating licensed library resources with sakai,” june , , www.dlib.indiana.edu/projects/sakai (accessed august , ). . university of rochester river campus libraries, “exten- sible catalog phase .” . susan gibbons, “library course management systems: an overview,” library technology reports , no. (may/june ): – . . marti a. hearst, “design recommendations for hier- archical faceted search interfaces,” august , http:// flamenco.berkeley.edu/papers/faceted-workshop .pdf (accessed august , ). . kristin antelman, emily lynema, and andrew k. pace, “toward a twenty-first century library catalog,” information technology and libraries , no. (september ): – . . “c ,” https://www.library.rochester.edu/c (accessed september , ). as of the time of this writing, the c prototype is available to the public. however, the prototype is no longer being developed, and this prototype may cease to be available at some point in the future. . charley pennell, “forward to the past: resurrecting faceted search @ ncsu libraries,” (powerpoint presenta- tion at the american library association annual conference, washington, d.c., june , ), www.lib.ncsu.edu/endeca/ presentations/ -facetedcatalogs-pennell.ppt (accessed august , ). . mary charles lasater, “authority control meets faceted browse: vanderbilt and primo,” (powerpoint presentation at the american library association annual conference, washington, d.c., june , ), www.ala.org/ala/lita/litamembership/ litaigs/authorityalcts/ annualfiles/marycharleslasater.ppt (accessed august , ). . casey bisson, “faceting and clustering: an implementa- tion report based on scriblio,” (powerpoint presentation at the american library association annual conference, washing- ton, d.c., june , ), http://oz.plymouth.edu/~cbisson/ presentations/alaannual_ - june .pdf (accessed august , ). . “subject access fields ( xx),” in marc concise format for bibliographic data ( ), www.loc.gov/marc/bibliographic/ ecbdsubj.html (accessed september , ). . pennell, “forward to the past: resurrecting faceted search@ ncsu libraries.” . “fast: faceted application of subject terminology,” www.oclc.org/research/projects/fast (accessed august , ). . ifla study group on the functional requirements for bibliographic records, functional requirements for bibliographic records. . ifla working group on functional requirements and numbering of authority records (franar), “functional requirements for authority data.” . library of congress, network development and marc standards office, “functional analysis of the marc bib- liographic and holding formats,” april , , www.loc. gov/marc/marc-functional-analysis/functional-analysis.html (accessed august , ); martha m. yee, “frbrization: a method for turning online public finding lists into online public catalogs,” information technology and libraries , no. (june ): – ; pat riva, “mapping marc linking entry fields to frbr and tillett’s taxonomy of bibliographic relation- ships,” library resources and technical services , no. (april ): – . . trond aalberg, “a process and tool for the conversion of marc records to a normalized frbr implementation,” in digital libraries: achievements, challenges and opportunities (berlin/heidelberg: springer, ), – ; christian monch and trond aalberg, “automatic conversion from marc to frbr,” in research and advanced technology for digital libraries (berlin/heidelberg: springer, ): – ; david mimno and gregory crane, “hierarchical catalog records: implementing a frbr catalog,” d-lib magazine , no. (october ), www .dlib.org/dlib/october /crane/ crane.html (accessed august , ). . trond aalberg, frank berg haugen, and ole husby, “a tool for converting from marc to frbr,” in research and advanced technology for digital libraries (berlin/heidelberg: springer, ), – ; “frbr work-set algorithm,” www .oclc.org/research/software/frbr/default.htm (accessed august , ); “xisbn (web service),” www.worldcat .org/affiliate/webservices/xisbn/app.jsp (accessed august , ). . for example, marc data may need to be augmented to extract data attributes related to frbr works and expressions that are not explicitly coded within a marc bibliographic record (such as a date associated with a work coded within a general note field); or to “sort out” the fields in a marc bibliographic record for a single resource that contains various works and/or expressions (e.g. ,a sound recording with multiple tracks), to associate the various fields (performer access points, analytical entries, subject headings, etc.) with the appropriate work or expression. . while the rit-developed tool is not publicly available at the time of this writing, it is our intent to post it to sourceforge (www.sourceforge.net) in the near future. the final report of the rit project is available at http://docushare.lib.rochester.edu/ docushare/dsweb/get/document- (accessed january , ). metadata to support next-generation library resource discovery | bowen . foster et al., “extensible catalog survey report.” . note the arrow pointing to the left in figure between the user environments and the metadata services hub. . jane greenberg and eva mendez, knitting the semantic web (binghamton, ny: haworth information press, ). this volume, co-published simultaneously as cataloging and clas- sification quarterly , no. / , contains a wealth of articles that explore the role that libraries can, and should, play in the devel- opment of the semantic web. . corey a. harper and barbara b. tillett explore various methods for making these controlled vocabularies available in “library of congress controlled vocabularies and their appli- cation to the semantic web,” cataloging and classification quar- terly , no. / ( ): . the development of skos (simple knowledge organization system), a semantic web language for representing controlled structured vocabularies, will also be valuable for xc. see alistair miles and jose r. perez-aguiera, “skos: simple knowledge organisation for the web,” catalog- ingand classification quarterly , no. / ( ). . marcus, “ethicshare planning phase final report.” . the talis platform provides another promising environ- ment for experimentation and development. see “talis platform: semantic web application platform,” talis, www.talis.com/ platform (accessed september , ). information technology and libraries | june author id box for column layout column title information technology and libraries | september communications james feher and tyler sondag administering an open-source wireless network this tutorial presents enhancements to an open-source wireless network dis- cussed in the june issue of ital that should reduce its administrative burden. in addition, it will demonstrate an open- source monitoring script written for the wireless network. as it has become increasingly impor- tant to provide wireless internet access for their patrons, libraries and colleges are almost expected to offer this service. inexpensive methods of providing wireless access—such as adding a commodity wireless access point to an existing network—can suffer from security issues, access by external entities, and bandwidth abuses. designs that address these issues often involve more costly pro- prietary hardware as well as expertise and effort that are often not readily available. a wireless network built with open-source software and com- modity hardware that addressed the cost, security, and equal access issues mentioned above was presented in the june issue of ital. this tutorial highlights enhancements to the pre- vious design that help to explain the technical hurdles in implementation, and includes a program that monitors the status of the various software and hardware components, helping to reduce the time required to administer the network. the wireless network presented requires several different pieces of soft- ware that must work together. because each of the required software programs are frequently updated, slight changes to the implementation may also be needed. a few issues that have arisen since the previous paper was written are addressed. a note is provided explaining the significance of setting the correct media access control (mac) address for the radius server and for wireless distribution system (wds) when configuring the system. in addition, in order to provide secure exchange of authentication credentials (username and password), the secure socket layer was used. a brief expla- nation of how to install a registered certificate on the gateway server is provided. lastly, a program that moni- tors the status of the network, provides a web page displaying the status of the various hardware and software components, and e-mails administra- tors with any changes to the network status—along with information on how this program is to be deployed within the network—is presented. configuration changes for previous design as new exploits are discovered and patched on a continual basis, any system should be regularly updated to insure that the most recent software is being used. the network design provided in the previous article used many different software components including, but not limited to: access point software openwrt—whiterussian rc dns cache dnsmasq v . gateway chillispot v . operating system fedora core radius server free radius v . . web caching server squid v . web server apache . . many of these components can be kept up-to-date by using the yellow dog updater, modified (yum). for example, to update a given package, with root access, at the command line enter: yum update packagename the yum command may also be used to update each package that has an available update by simply removing the package name from the yum update command and entering the following: yum update yum may also be used to upgrade the entire operating system. keep in mind that with any change in software, the configuration of any particular package may change as well. for example, the newest version of squid is currently . . appendix d in the previous paper explained how to allow transparent relay of web requests so that client browsers did not have to be reconfigured. so, while version . required four changes to allow the transparent relay, the current version—found in appendix a—requires only one. in addition to changes in software, occasionally even entire websites move, as happened with chillispot. another change involved the con- figuration of the linksys wrt gs access points. the newer versions of this access point/router sold by linksys have half the flash memory and half of the ram of the older ver- sions. while the newer versions of the linksys wrt gs can be flashed with custom firmware, the firmware that will fit on the newer unit lacks all the capability of the standard firmware. given this, those wishing to imple- ment such a wireless network should investigate the capability of models to be deployed, as well as the version numbers for the access points chosen. the current version of the linksys wrt gl and wrtsl gs units retain enough flash memory and ram to be updated with the standard firm- ware mentioned in the previous article. james feher (jdfeher@mckendree.edu) is associate professor of computer science and computer information systems at mckendree university, lebanon, illinois. tyler sondag (sondag@cs.iastate.edu) is a phd candidate in computer science at iowa state university, ames. introducing zoomify image | smith administering an open-source wireless network | feher and sondag in addition, the procedure for upgrad- ing the firmware for the wrtsl gs is simpler than the procedure outlined in appendix i of the previous paper. the factory-installed firmware on version . can be flashed directly using the web interface provided by linksys. so, while this tutorial and the previous paper outline the design of a network, the administrator will need to be vigilant in updating the packages used and keep in mind that the configuration specifications may also change with those updates. the administrator for the network must also investigate the capability of the standard hardware used to insure that it retains the functionality required for the system. choosing the correct mac address for the access point the access points used will have more than one interface and as such more than one mac address. when enter- ing the mac address of a given access point into either the users file for the radius server or the access points that use the wds, use the mac address associated with the wireless interface. using the incorrect mac address will result in problems when communicat- ing with the various access points. for the radius server, the access point will not get the correct ip address, which will prohibit the possibility of remotely administering the unit. incorrect mac addresses that are used for the wds settings will cause even worse problems, as the unit will not be able to relay data from users who connect to this access point. installation of a registered ssl certificate as users are required to enter their authentication credentials to gain access to the internet, the exchange of this data is encrypted using the secure socket layer. while administrators can self-sign the certificates used for their web servers, it is recommended that a registered certificate be obtained and installed for the system. this can help prevent common attacks and has the added benefit of eliminating warnings for the client browsers when they detect unregistered certificates being used by the ssl. a search of “ssl certificate” will yield any number of commercial vendors from which a certificate can be obtained. generally the installation of a certificate is fairly straightforward. the openssl com- mand line utility can be used to gener- ate a ssl key and certificate signing request (csr). once the csr is generated, pick a vendor/certificate authority who can sign your key. it should be noted that the design presented required the authentica- tion gateway to be behind the main router. this required a certificate to be signed for a server within an intranet that does not have a fully qualified domain name. so, when generating the ssl key and csr, make sure to use gatewayhostname.localnet as the common name of your server. of course, gatewayhostname is whatever you choose as the name of your gateway host. the term localnet is used to refer to the server existing within an intranet. then make sure to place an entry for gatewayhostname.localnet into the hosts file of the server that is pro- viding domain name service for your network. an example entry for the hosts file which is in the /etc directory of a standard fedora core installation is found in appendix b. monitoring script for wireless network as the wireless network has many separate hardware and software components, many possible points of failure exist for the system. the script from appendix c, which was written in perl, uses ping to test if each access point is still connected to the network and nmap to test whether the port associated with a given net- work service is still available. this program can be run manually or, even better, run automatically through the unix cron utility to update a webpage that displays the current state of all the network components. the web- page generated by this script for the mckendree college wireless network may be found at http://lance.mcken- dree.edu/cgi-bin/wireless/status.cgi. (additionally, a sample of this page is available as a figure in appendix d.) this script actually contains a script within a script. the main script must be run on the gateway machine, chilli on the diagram in appendix e, as only this machine has access to ping the access points. when the script determines that an access point or daemon is down, it will e-mail the system administrator. when an access point is down, in addition to sending the system administrator an e-mail, it can also send notification to an e-mail address associated with that device. this allows for someone other than the system administrator—who may have closer physical access to the unit—to check the access point on behalf of the administrators for simple issues, such as an access point losing power. this script then generates another cgi script that can be transmitted to an external server that can be reached from anywhere on the internet. in this case, this generated script can be run as a web-based application or by the system itself using the cron utility. if run as by the cron daemon, it will also e-mail the administrators if the script has not been updated recently. the script requires the use of several perl modules that will need to be installed. n expect n mail::mailer n net::ping the script has been released using the gnu general public license, information technology and libraries | june information technology and libraries | september version (gpl). the first portion of the script contains a reference to the gpl, followed by a brief explanation of the script as well as a set of parameters that should be changed to fit the specifications of the network designed. conclusion administrators should be vigilant in updating the entire system to assure security, keeping in mind that new versions of software or hardware may necessitate changes in the overall con- figuration of the system. in addition, while the monitoring script provides a useful aid in monitoring the net- work, it could be further expanded to include a more comprehensive review of level of use for various access points by the different users. it is felt that this would be best done through a database, which would require a higher level of administrative effort. a brief frequently asked questions list along with the script and link to the code for the script can be found at http://lance.mckendree.edu/csi/ wirelessfaq.html. references . sondag, tyler and james feher, “open source wifi hotspot implementa- tion,” information technology and libraries , no. : – , http://ala.org/ala/lita/ litapublications/ital/ / jun/ toc.cfm (accessed july , ). . linux@duke, “yum: yellow dog updater, modified,” http://linux.duke .edu/projects/yum (accessed july , ) . upgrading fedora using yum fre- quently asked questions, http://fedora p r o j e c t . o r g / w i k i / yu m u p g r a d e f a q (accessed mar. , ). . chillispot—open source wire- less lan access point controller, “spice up your hotspot with chilli,” www .chillispot.info/ (accessed may , ). . openwrtdocs/hardware/linksys /wrt gs—openwrt, http://wiki.open wrt.org/openwrtdocs/hardware/link sys/wrt gs (accessed july , ). . bitsum technologies wiki— wrt g cfe, http://bitsum.com/ openwiking/owbase/ow.asp?wrt g _ cfe (accessed july , ). . openwrtdocs/hardware/link- sys/wrtsl gs—openwrt, http:// wiki.openwrt.org/openwrtdocs/hard ware/linksys/wrtsl gs (accessed july , ). . o p e n wr t d o c s / w h i t e r u s s i a n / configuration, wireless distribution sys- tem (wds)/repeater/bridge. http:// wiki.openwrt.org/openwrtdocs/white russian/configuration (accessed july , ). . viega, john, matt messier, and pravir chandra, network security with openssl cryptography for secure commu- nications. (sebastopol, calif.: o’reilly and associates, ). . generating a key pair and csr for an apache server with modssl. www .verisign.com/support/tlc/csr/modssl/ v .html (accessed feb. , ). . wall, larry, tom christiansen, and randal schwartz, programming perl, third edition (sebastopol, calif.: o’reilly and associates). . nmap—free security scanner for network exploration and security audits. http://insecure.org/nmap/ (accessed feb. , ). . gnu general public license ver- sion , june . www.gnu.org/licenses/ gpl.txt. appendix a. squid configuration changes # changes made to squid.conf # lines needed for squid . #httpd_accel_port #httpd_accel_host virtual #httpd_accel_with_proxy on #httpd_accel_uses_host_header on # # one line needed in version . http_port transparent appendix b. /etc/hosts entry on marla for localnet entry . . . marla localhost.localdomain localhost . . . bob . . . lance.mckendree.edu lance # next line is for the ssl certificate to work properly . . . chilli.localnet chilli introducing zoomify image | smith administering an open-source wireless network | feher and sondag appendix c. monitoring script #!/usr/bin/perl ######################################################### # code released / / under: # # the gnu general public license, version # # http://www.gnu.org/licenses/gpl.txt # # # # it is recommended that this script is run as a cron # # job frequently to find changes in the network. this # # script will check the status of the wireless access # # points/routers as well as the daemons necessary to # # run the network. it will then output the results to # # another perl file that is copied to a remote # # webserver. when the script observes a change in the # # availability of any access point or daemon, email # # will be sent to the specified administrator # # address(es). the option exists to send an email to # # to an additional person for each access point. # # # # additionally, the output file on the remote webserver # # will check when it was last updated, if that script # # is run from the command line or via cron. if it has # # not been updated for a specified number of minutes, # # it will send an email to the administrator. it is # # also recommended that this output script be run as a # # cron jobr. this output script can also be executed # # as a cgi program to generate a display of network # # status. # ######################################################### use strict; use expect(); # needed to scp to webserver use mail::mailer; # needed to send emails if outages use net::ping; # needed to check the status of aps #variables for webserver to host status page’s my $webservuname = “username”; my $webservpass = “password”; my $webservurl = “lance.mckendree.edu”; my $webservtarg = “/var/www/cgi-bin/wireless/”; my $weboutputurl = “http://lance.mckendree.edu/cgi-bin/wireless/status.cgi”; my $instname = “mckendree college”; #default background color of the status page my $defbgcolor = “# ”; # if the page on the webserver has not been updated # in $updatemin minutes send an email that the service # is down (set to =~ *crontime) my $updatemin = ; #email address errors will be sent to my $fromemail = ‘admin @email.com’; my $toemail = information technology and libraries | june information technology and libraries | september ‘admin @email.com, admin @email.com’; #file where errors will be stored on remote host my $logfilename = “/tmp/wireleslog.txt”; #hash for routers/ap’s #location is displayed on the webpage and in status emails #owner - changes in status regarding this ap are sent to # this address as well (optional) my %iptoloc = ( “ . . . ” => { “location” => “clark ”, “owner” => ‘’}, “ . . . ” => { “location” => “clark a”, “owner” => ‘apuser @email.com’}, “ . . . ” => { “location” => “pac lounge”, “owner” => ‘apuser @email.com’}, “ . . . ” => { “location” => “library main”, “owner” => ‘apuser @email.com’}, “ . . . ” => { “location” => “library upper”, “owner” => ‘’}, “ . . . ” => { “location” => “library lower”, “owner” => ‘’}, “ . . . ” => { “location” => “carnegie”, “owner” => ‘apuser @email.com’}); #hash for daemons my %daemons = ( “dnsmasq - dns server” => { “ip_addr” =>” . . . ”, “port” =>” ”, “proto” =>”tcp”}, “radius - authenticate” => { “ip_addr” =>” . . . ”, “port” =>” ”, “proto” =>”udp”}, “chilli - capt. portal” => { “ip_addr” =>” . . . ”, “port” =>” ”, “proto” =>”local”}, “squid - web cache” => { “ip_addr” =>” . . . ”, “port” =>” ”, “proto” =>”tcp”}, “apache - web server” => { “ip_addr” =>” . . . ”, “port” =>” ”, “proto” =>”tcp”}); introducing zoomify image | smith administering an open-source wireless network | feher and sondag ######################################################## # # # no changes need to be made to the following code # # # ######################################################## # get the current time my $currenttime = scalar localtime(); my $starttime = time(); # open old output status script to get previous status’ open(old, “status.cgi”); my @tmpoldstatfile = ; my $oldstatfile = join(“”, @tmpoldstatfile); # check routers/ap’s using ping my $diff = ‘’; my $allrouterstat; foreach my $host (sort keys %iptoloc){ my $p = net::ping->new(); my $pingresult = $p->ping($host); if(!$pingresult){ sleep ; $pingresult = $p->ping($host); } my $thislaststat = ( $oldstatfile =~ m/$iptoloc{$host}{location}<\/td>close(); } #check the status of each daemon my $alldaemonstat =’’; foreach my $i (sort keys %daemons){ my $thislaststat = ( $oldstatfile =~ m/$i<\/td> (\$lasttime + ( * $updatemin))){ \$systemstatus = “#ff ”; \$message = “status update failed”; } # if this is cron running the script if (\$currentuser =~ “$webservuname”){ # send email if status is down & logfile doesn’t exist &sendemail() if ( (\$systemstatus =~ “#ff ”) && !(-e “$logfilename”) ); # delete log file if everything is up unlink(“$logfilename”) if ( (!(\$systemstatus =~ “#ff ”)) && (-e “$logfilename”) ); } #else apache is accessing the page (its a web request) else{ #print the page print header(); ############################ # start of html output # ############################ print < $instname wireless status introducing zoomify image | smith administering an open-source wireless network | feher and sondag $instname wireless status \$message $allrouterstat
access point status

$alldaemonstat
daemon status


last updated $currenttime
web_output ########################## # end of html output # ########################## }#end else sub sendemail { my \$mailer = mail::mailer->new(“sendmail”); \$mailer->open({from => ‘$fromemail’, information technology and libraries | june information technology and libraries | september to => [\$toemail], subject => “wireless problem”}); my \$message = “the wireless system has failed to “ .”it’s status.\n\n$weboutputurl\n”; print \$mailer \$message; \$mailer->close(); open(file, “>>$logfilename”); print file “failed to update system.”; close(file); } output_file_for_remote_host ######################################################## # end of script output block # ######################################################## #write output code to the file my $perloutputfile = “status.cgi”; open (out, “>$perloutputfile”); print out $perloutput; close (out); chmod , $perloutputfile; #send email is necessary &sendemail($diff, $weboutputurl, $fromemail, $toemail) if ($diff); #send perl file to webserver &scpfile($perloutputfile, $webservuname, $webservpass, $webservurl, $webservtarg); ################################################ # # # end main code block, start functions # # # ################################################ # given the name and status of something (ap or # daemon), this returns a string for the table # row for displaying the status of the ap/daemon sub printstatus { my ($service, $status, $oldstatus, $owner, $toemail,$oldstatusfile, $currenttime ) = @_; my $msg = “”; my $statusline = “\n $serviceup”; introducing zoomify image | smith administering an open-source wireless network | feher and sondag # if last two status’ were down if ($oldstatusfile =~ m/\($service\)- --->/){ $msg = “$service back up at $currenttime\n”; # if service has owner & not already in mail list, # add owner to mail list $toemail .= “, \’$owner\’” if ($owner && (!($toemail =~ $owner))); } } #else current status is down else{ $statusline .= “down\”>down”; # if last status was down & before that status was up if ($oldstatusfile =~ m/\($service\)- - -->/){ $msg = “$service down at $currenttime\n”; # if service has owner & not already in mail list, # add owner to mail list $toemail .= “, \’$owner\’” if ($owner && (!($toemail =~ $owner))); } } $statusline .= “”; return ($statusline, $toemail, $msg); }#end printstatus function # checks the status for the given daemon # takes in ip, port to check, daemon name, and protocol # (tcp/udp). if given port= it checks for local daemon sub checkdaemon { my ($ip, $port, $daemon, $proto) = @_; my $dstat = ; if ($proto !~ /local/){ #su checks for udp ports my $com = ($proto =~ “tcp”) ? (“nmap -p $port $ip | grep $port”) : (“nmap -su -p $port $ip | grep $port”); open(tmp, “$com|”); my $comout = ; close(tmp); if ($comout =~ /open/){ $dstat = ; #if port is open, status is up } } else{ $daemon =~ s/ +.*//g; #\l lowercases the first letter of $daemon my $com = “which \l$daemon”; open(tmp, “$com|”); my $comout = ; close(tmp); $com = “ps aux | awk ‘{print \$ }’ | grep $comout”; open(tmp, “$com|”); $comout = ; close(tmp); $dstat = if ($comout); information technology and libraries | june information technology and libraries | september } return $dstat; } # end checkdaemon function # send the output perl status file to the webserver sub scpfile { my ($filepath, $webservuname, $webservpass, $webservurl, $webservtarg ) = @_; my $command = “scp $filepath $webservuname” .”\@$webservurl:$webservtarg”; my $exp = expect->spawn ($command); # the first argument “ ” may need to be adjusted # if your system has very high latency my $ret = $exp ->expect( , “word:”); print $exp “$webservpass\r”; my $ret = $exp ->expect(undef); $exp ->close(); } # end scpfile function # send an email to the admin & append error to log file sub sendemail { my ($errorlist, $weboutputurl, $fromemail, $toaddresses ) = @_; my $mailer = mail::mailer->new(“sendmail”); $mailer->open({from => “$fromemail”, to => [$toaddresses], subject => “wireless problem”}); $errorlist .= “\n\n$weboutputurl”; print $mailer $errorlist; $mailer->close(); } # end sendemail function appendix d. script output page appendix e. diagram of network lita cover , cover , cover index to advertisers information technology and libraries | june preparing locally encoded electronic finding aid inventories for union environments: a publishing model for encoded archival description author id (to come) plato l. smith ii this paper will briefly discuss encoded archival descrip- tion (ead) finding aids, the workflow and process involved in encoding finding aids using ead metadata standard, our institution’s current publishing model for ead finding aids, current ead metadata enhancement, and new developments in our publishing model for ead finding aids at florida state university libraries. for brevity and within the scope of this paper, fsu libraries will be referred to as fsu, electronic ead finding and/ or archival finding aid will be referred as ead or eads, and locally encoded electronic ead finding aids invento- ries will be referred to as eads @ fsu. n what is an ead finding aid? many scholars, researchers, and learning and schol- arly communities are unaware of the existence of rare, historic, and scholarly primary source materials such as inventories, registers, indexes, archival documents, papers, and manuscripts located within institutions’ col- lections/holdings, particularly special collections and archives. a finding aid—a document providing informa- tion on the scope, contents, and locations of collections/ holdings—serves as both an information provider and guide for scholars, researchers, and learning and schol- arly communities, directing them to the exact locations of rare, historic, and scholarly primary source materi- als within institutions’ collections/holdings, particularly noncirculating and rare materials. the development of the finding aid led to the institution of an encoding and markup language that was software/hardware indepen- dent, flexible, extensible, and allowed online presentation on the world wide web. in order to provide logical structure, content pre- sentation, and hierarchical navigation, as well as to facilitate internet access of finding aids, the university of california–berkeley library in initiated a coop- erative project that would later give rise to development of the nonproprietary sgml-based, xml-compliant, machine-readable markup language encoding finding aid standard, encoded archival description (ead) docu- ment type definition (dtd) (loc, a). thus, an ead finding aid is a finding aid that has been encoded using encoded archival description and which should be validated against an ead dtd. the ead xml that pro- duces the ead finding aid via an extensible style sheet language (xsl) should be checked for well-formed-ness via an xml validator (i.e. xml spy, oxygen, etc.) to ensure proper nesting of ead metadata elements “the ead document type definition (dtd) is a stan- dard for encoding archival finding aids using extensible markup language (xml)” (loc, c). an ead finding aid includes descriptive and generic elements along with attribute tags to provide descriptive information about the finding aid itself, such as title, compiler, compilation date, and the archival material such as collection, record group, series, or container list. florida state university libraries has been creating locally encoded electronic encoded archival descrip- tion (ead) finding aids using a note tab light text editor template and locally developed xsl style sheets to generate multiple ead manifestations in html, pdf, and xml formats online for over two years. the formal ead encoding descriptions and guidelines are developed with strict adherence to the best practice guidelines for the implementation of ead version in florida institutions (fcla, ), manuscript processing reference manual (altman & nemmers, ), and ead version . an ead note tab light template is used to encode findings down to the collection level and cre- ate ead xml files. the ead xml files are tranformed through xsl stylesheets to create ead finding aids for select special collections. n ead workflow, processes, and publishing model the certified archivist and staff in special collections and a graduate assistant in the digital library center encode finding aids in ead metadata standard using an ead clip and ead template library in note tab light text editor via data entry input for the various descriptive, administrative, generic elements, and attribute metadata element tags to generate ead xml files. the ead xml files are then checked for validity and well-formed-ness using xml spy . currently, ead finding aids are encoded down to the folder level, but recent florida heritage project – grant funding has allowed selected special collections finding aids to be encoded down to the item level. currently, we use two xsl style sheets, ead html.xsl and ead pdf.xsl, to generate html and pdf formats, and simply display the raw xml as part of rendering ead finding aids as html, pdf, and xml and present- ing these manifestations to researchers and end users. the ead html.xsl style sheet used to generate the html versions was developed with specifications such as use of fsu seal, color, and display with input from the special collections department head. the ead pdf.xsl style sheet used to generate pdf versions uses xsl-fo (formatting plato l. smith ii (psmithii@fsu.edu) is digital initiatives librarian at florida state university libraries, tallahassee. preparing locally encoded electronic finding aid inventories for union environments | smith object), and was also developed with specifications for layout and design input from the special collections department head. the html versions are generated using xml spy home edition with built-in xslt, and the pdf versions are generated using apache formatting object processor (fop) software from the command line. ead finding aids, eads @ fsu, are available in html, pdf, and xml formats (see figure ). the style sheets used, ead authoring software, and eads @ fsu origi- nal site are available via www.lib.fsu.edu/dlmc/dlc/ findingaids. n enriching ead metadata as ead standards and developments in the archival community advance, we had to begin a way of enrich- ing our ead metadata to prepare our locally encoded ead finding aids for future union catalog searching and opac access. the first step toward enriching the metadata of our ead finding aids was to use rlg ead report card (oclc, ) on one of our ead finding aids. the test resulted in the display of missing required (req), mandatory (m), mandatory if applicable (ma), recommended (rec), optional (opt), and encoding analogs (relatedencoding and encodinganalog attri- butes) metadata elements (see figure ). the second test involved reference online archive of california best practices guidelines (oac bpg), specifically appendix b (cdl, , ¶ ), to create a formal public identifier (fpi) for our ead finding aids and make the ead fpis describing archives content standards (dacs)–compliant. this second test resulted in the creation of our very first dacs– compliant ead formal public identifier. example: ftasu . xml the rlg ead report card and appendix b of oac bpg together helped us modify our ead finding aid encoding template and work- flow to enrich the ead document identifier metadata tag element, include miss- ing mandatory ead metadata elements, and develop fpis for all of our ead finding aids. prior to recent new developments in the publishing model of ead finding aids at fsu libraries, the ead finding aids in our eads @ fsu inventories could not be easily found using traditional web search engines, were part of the so-called “deep web,” (prom & habing, ) and were “unidimensional in that they [were] based upon the assumption that there [was] an object in a library and there [was] a descriptive surrogate for that object, the cataloging record” (hensen, ). ead finding aids in our eads @ fsu inventories did not have a descriptive surrogate catalog record and lacked the relevant related encoding and analog metadata ele- ments within the ead metadata with which to facilitate “metadata crosswalks”—mapping one metadata stan- dard with another metadata standard to facilitate cross- searching. “to make the metadata in ead instance as robust as possible, and to allow for crosswalks to other encoding schemes, we mandate the inclusion of the relat- edencoding and encodinganalog attributes in both the and segments” (meissner, et al., ). incorporating an ead quality checking tool such as rlg bpg and ead compliance such as dacs when figure . ead finding aids in html, pdf, and xml format figure . rlg ead report card of xml ead file information technology and libraries | june authoring eads, will assist in improving ead encoding and ead finding aids publishing model. n some key issues with creating and managing ead finding aids one of the major issues with creating and managing ead finding aids is the set of rules used for describing papers, manuscripts, and archival documents. the former set of rules used for providing consistent descriptions and anglo-american cataloging rules (aacr) bibliographic catalog compliance for papers, manuscripts, and archi- val documents down to collection level was archives, personal papers, and manuscripts (appm), which was complied by steven l. hensen and published by the library of congress in . however, the need for more description granularity down to the item level, enhanced bibliographic catalog specificity, marc and ead meta- data standards implementations and metadata standards crosswalks, and inclusion of descriptors of archival material types beyond personal papers and manuscripts prompted the development of describing archives: a content standard (dacs), published in with the second edition published in . “dacs [u.s. imple- mentation of international standard for the description of archival materials and their creators] is an output-neutral set of rules for describing archives, personal papers, and manuscripts collections, and can be applied to all mate- rial types ”(pearce-moses, ). some international standards for describing archival materials are general international standard archival description isad(g) and international standard archival authority record for corporate bodies, persons, and families [isaar(cpf)]. other issues with creating and managing ead find- ing aids include (list not exhaustive): . online presentation of finding aids . exposing finding aids electronically for searching . provision of a search interface to search finding aids . online public access catalog record (marc) and link to finding aids . finding aids linked to digitized content of collections eads @ fsu exist in html for online presenta- tion, pdf for printing, and xml for exporting, which allow researchers greater flexibility and options in the information-gathering and research processes and have improved the way archivists communicated guides to archival collections with researchers as opposed to paper finding aids physically housed within institutions. eads @ fsu have existed online in html, pdf, and xml formats for two years in a static html document and then moved to drupal (mysql database with php) for about one year, which improved online maintenance but not researcher functionality. however, the purchase and upgrade of a digital content management system marked a huge advancement in the development of our ead finding aids implementation and thus resolutions to issues numbers – . researchers now have a single-point search interface to search eads @ fsu across all our digital collections/ institutional repository (see figure ); the ability to search within the finding aids via full-text indexing of pdfs; the option of brief (thumb- nails with ead, htm, pdf, and xml manifes- tation icons), table (title, creator, and identifier), and full (complete ead finding aid dc record with manifestations) views of search results, which provides different levels of exposures of ead finding aids; and the ability to save/e-mail search results. future initiatives are underway to enhance eads @ fsu implementation via the creation of ead marc records through dublin core to marc metadata crosswalk, to deep link to ead finding aids via field in marc records, and to begin digitizing and linking to ead finding aids archival content via digital archival object ead element. is “linking element that uses the attributes entityref or href to connect the finding aid information to elec- tronic representations of the described materi- als. the and elements allow the content of an archival collection or record figure . online search gui for ead finding aids and digital collections within ir preparing locally encoded electronic finding aid inventories for union environments | smith group to be incorpo- rated in the finding aid” (loc, b). we have opted to create basic dublin core records of ead finding aids based on the information in the ead finding aids descriptive summary (front matter) first and then crosswalk to marc, but are cogni- zant that this current workflow is subject to change in the pur- suit of advancement. however, we are seeking ways to improve the ead workflow and ead marc record creation through more communication and future collaboration with the fsu libraries cataloging department. n number of finding aids and percent of eads @ fsu as of february , , we had collections with finding aids in which finding aids are electronic and encoded in html ( percent of total finding aids). from the electronic finding aids, are available as html, pdf, and xml finding aids ( percent of electronic find- ing aids are eads @ fsu). however, we currently have ead finding aids available online in html, pdf, and xml formats. n new developments in publishing eads @ fsu current eads @ fsu include the recommendations from test and test (rlg bpg and dacs compliance) which were discussed earlier and the digital content manage- ment system (i.e. digitool) creates a descriptive digital surrogate of the ead objects in the form of brief and basic dublin core metadata records for each ead finding aid along with multiple ead manifestations (see figure ). we have successfully built and launched our first new digital collection, fsu special collections ead inventories, in digitool . as part of fsu libraries dlc digital repository (http://digitool .lib.fsu.edu/r/), a relational database digital content management system (dcms). digitool has an oracle i relational database management system backend, searchable web-based gui, a default ead style sheet that allows full-text searching of eads, supports marc, dc, mets metadata standards, jpeg (built in tools for images and thumbnails) as well as z . and oai protocols which will enable resource discovery and exposing of eads @ fsu. you can visit fsu special collections ead finding aids inventories at http://digitool .lib.fsu.edu/r/? func=collections-result&collection_id= . n national, international, and regional aggregation of finding aids initiatives rlg’s archivegrid (http://archivegrid.org/web/index. jsp) is an international, cross-institutional search consti- tuting the aggregation of primary source archival materi- als of more than , research libraries, museums, and archives with a single-point interface to search archival collections from across research institutions. other inter- national, cross-institutional searches of aggregated archi- val collections are: n intute: arts& humanities in the united kingdom www.intute.ac.uk/artsandhumanities/ cgi-bin/browse.pl?id= (international guide to subcategories of archival materials) n archives made easy www.archivesmade easy.org (guide to archives by country) there are also some regional initiatives, which pro- vide cross-institutional search of aggregations of finding aids: n publication of archival library and museum materials (palmm) http://palmm.fcla.edu (cross- figure . ead finding aids in ead (default), html, pdf, and xml manifestations information technology and libraries | june institutional searches in fl - fsu participates, fl) n virginia heritage: guides to manuscript and archival collections in virginia http://ead.lib .virginia.edu/vivaead/ (cross-institutional searches in virginia) n texas archival resources online www.lib.utexas. edu/taro/ (cross-institutional searches in texas) n online archive of new mexico http://elibrary .unm.edu/oanm/ (cross-institutional searches in new mexico) awareness of regional, national, and international aggregation of finding aids initiatives and engagement in regional aggregation of finding aids will enable a consis- tent advancement in the development and implementa- tion of eads @ fsu. acknowledgments fsu libraries digital library center and special collections department, florida heritage project funding (fcla), chuck f. thomas (fcla), and robert mcdonald (sdsc) assisted in the development, implementation, and success of eads at fsu. references altman, b. & nemmers, j. ( ). manuscripts processing ref- erence manual. florida state university special collections. california digital library (cdl). ( ). oac best practice guidelines for encoded archival description, appendix b. for- mal public identifiers for finding aids. retrieved october , from www.cdlib.org/inside/diglib/guidelines/bpgead/ bpgead_app.html#d e . digital library center, florida state university libraries. ( ). fsu special collections ead finding aids inventories. retrieved january , from http://digitool .lib.fsu.edu/ r/?func=collections-result&collection_id= . florida center of library automation (fcla). ( ). palmm: publication of archival library and museum materials, archival collections. retrieved january , from http://palmm.fcla .edu. florida center for library automation (fcla). ( ). best practice guidelines for the implementaton of ead version in florida institutions. (john nemmers, ed.). accessed april , , at www.fcla.edu/dlini/openingarchives/new/ floridaeadguidelines.pdf fox, m. ( ). the ead cookbook — edition.chicago: the society of american archivists. retrieved october , from www.archivists.org/saagroups/ead/ead cookbook .html. hensen, s. l. ( ). nistf ii and ead: the evolution of archi- val description. encoded archival description: context, theory, and case studies (pp. – ). chicago: the society of american archivsits library of congress (loc). ( a). development of the encoded archival description dtd. retrieved october , from www.loc.gov/ead/eaddev.html. library of congress (loc). ( b). digital archival object— encoded archival description tag library—version . retrieved january , from www.loc.gov/ead/tglib. library of congress (loc). ( c). encoded archival descrip- tion —version official site. etd dtd version . retrieved april , from www.loc.gov/ead/ead a.html. meissner, d., kinney, g., lacy, m., nelson, n., proffitt, m., rinehart, r., ruddy, d., stockling, b., webb, m., & young, t. ( ). rlg best practices guidelines for encoded archival description (pp. - ). mountain view: rlg. retrieved january , from www.rlg.org/en/pdfs/bpg.pdf. national library of australia. ( ). use of encoded archi- val description (ead) for manuscript collection retrieved january , from www.nla.gov.au/initiatives/ead/eadintro .html. oclc. ( ). archivegrid—open the door to history. retrieved january , from http://archivegrid.org/web. oclc. ( ). ead report card. retrieved april , www.oclc.org/programs/ourwork/past/ead/reportcard .htm. pearce-moses, r. ( ). a glossary of archival and records terminology. chicago: society of american archivists. retrieved january , from www.archivists.org/glossary/index.asp. prom, c. j. & habing, t. g. ( ). using the open archives ini- tiative protocols with ead . paper preserted at the international conference on digital libraries proceedings of the nd acm/ieee-cs joint conference on digital libraries. portland, oregan, usa, july - , . retrieved october , from http://portal.acm .org/citation.cfm?doid= . . reese, t. ( ). building lite-weight ead repositories,. paper presented in the international conference on digital libraries proceedings of the th acm/ieee-cs joint conference on digital libraries. new york: acm. retrieved january , from http://doi.acm.org/ . / . . special collections department, university of virginia. ( ). virginia heritage guides to manuscripts and archival collections in virginia. retrieved january , from http://ead.lib.virginia .edu/vivaead/. thomas, c., et al. ( ). best practices guidelines for the implementation of ead version in florida institutions. florida state university special collections. university of texas libraries, university of texas at austin. (unknown). texas archival resources online (taro). retrieved january , from www.lib.utexas.edu/taro. information technology and libraries | september employing virtualization in library computing: use cases and lessons learned arwen hutt, michael stuart, daniel suchy, and bradley d. westbrook this paper provides a broad overview of virtualization technology and describes several examples of its use at the university of california, san diego libraries. libraries can leverage virtualization to address many long-standing library computing challenges, but careful planning is needed to determine if this technology is the right solution for a specific need. this paper outlines both technical and usability considerations, and concludes with a discussion of potential enterprise impacts on the library infrastructure. o perating system virtualization, herein referred to simply as “virtualization,” is a powerful and highly adaptable solution to several library technology challenges, such as managing computer labs, automat- ing cataloging and other procedures, and demonstrating new library services. virtualization has been used in one manner or another for decades, but it is only within the last few years that this technology has made significant inroads into library environments. virtualization technol- ogy is not without its drawbacks, however. libraries need to assess their needs, as well as the resources required for virtualization, before embarking on large-scale imple- mentations. this paper provides a broad overview of virtualization technology and explains its benefits and drawbacks by describing some of the ways virtualization has been used at the university of california, san diego (ucsd) libraries. n virtualization overview virtualization is used to partition the physical resources (processor, hard drive, network card, etc.) of one com- puter to run one or more instances of concurrent, but not necessarily identical, operating systems (oss). traditionally only one instance of an operating system, such as microsoft windows, can be used at any one time. when an operating system is virtualized—creating a vir- tual machine (vm)—the vm communicates through vir- tualization middleware to the hardware or host operating system. this middleware also provides a consistent set of virtual hardware drivers that are transparent to the end- user and to the physical hardware. this allows the virtual machine to be used in a variety of heterogeneous envi- ronments without the need to reconfigure or install new drivers. with the majority of hardware and compatibility requirements resolved, the computer becomes simply a physical presentation medium for a vm. n two approaches to virtualization: host-based vs. hypervisor virtualization can be implemented using type or type hypervisor architectures. a type hypervisor (figure ), commonly referred to as “host-based virtualization,” requires an os such as microsoft windows xp to host a “guest” operating system like linux or even another ver- sion of windows. in this configuration, the host os treats the vm like any other application. host-based virtualiza- tion products are often intended to be used by a single user on workstation-class hardware. in the type hypervisor architecture (figure ), com- monly referred to as “hypervisor-based virtualization,” the virtualization middleware interacts with the comput- er’s physical resources without the need of a host operat- ing system. such systems are usually intended for use by multiple users with the vms accessed over the network. realizing the full benefits of this approach requires a con- siderable resource commitment for both enterprise-class server hardware and information technology (it) staff. n use cases archivists’ toolkit the archivists’ toolkit (at) project is a collabora- tion of the ucsd libraries, the new york university libraries, and the five colleges libraries (amherst college, hampshire college, mt. holyoke college, smith college, and university of massechusetts, amherst) and is funded by the andrew w. mellon foundation. the at is an open-source archival data management system that provides broad, integrated support for the management of archives. it consists of a java client that connects to a relational database back-end (mysql, mssql, or oracle). the database can be implemented on a networked server or a single workstation. since its initial release in december , the at has sparked a great deal of interest and rapid uptake of the appli- cation within the archival community. this growing interest has, in turn, created an increased demand for demonstrations of the product, workshops and training, and simpler methods for distributing the application. (of the use cases described here, the two for the at arwen hutt (ahutt@ucsd.edu) is metadata specialist, michael stuart (mstuart@ucsd.edu) is information technology analyst, daniel suchy (dsuchy@ucsd.edu) is public services technology analyst, and bradley d. westbrook (bradw@library.ucsd.edu) is metadata librarian and digital archivist, university of california, san diego libraries. employing virtualization in library computing | hutt et al. distribution and laptop classroom are exploratory, whereas the rest are in production.) at workshops the society of american archivists sponsors a two-day at workshop occurring on multiple dates at sev- eral locations. in addition, the at team provides one- and two-day workshops to different institu- tional audiences. at workshops are designed to give participants a hands-on experience using the at application. accomplishing this effectively requires, at the mini- mum, supplying all participants with identical but separate data- bases so that participants can com- plete the same learning exercises simultaneously and independently without concern for working in each other’s space. in addition, an ideal configuration would reduce the workload of the instructors, freeing them from having to set up the at instructional database onsite for each workshop. for these workshops we needed to do the following: n provide identical but sepa- rate databases and database content for all workshop attendees n create an easily reproduc- ible installation and setup for workshops by prepar- ing and populating the at instructional database in advance virtualization allows the at workshop instructors to predefine the workstation configuration, including the installation and pop- ulation of the at databases, prior to arriving at the workshop site. to accomplish this we developed a workshop vm configuration with mysql and the at client installed within a linux ubuntu os. the workshop instructors then built the at vm with the data they require for the workshop. the at client and database are loaded on a dvd or flash drive and shipped to the classroom managers at the workshop sites, who then need only to install a copy of the vm and the freely available vmplayer software (necessary to launch the at vm) onto each workstation in the classroom. the at vm, once built, can be used many times both for multiple workstations in a classroom as well as for multiple work- shops at different times and locations. this implementation has worked very well, saving both time and effort for the instructors and classroom support staff by reducing the time and communication figure . a type hypervisor (host-based) implementation figure . a type hypervisor-based implementation information technology and libraries | september necessary for deploying and reconfiguring the vm. it also reduces the chances that there will be an unexpected conflict between the application and the host worksta- tion’s configuration. but the method is not perfect. more than anything else, licensing costs motivated us to choose linux as the operating system instead of a proprietary os such as windows. this reduces the cost of using the vm, but it also requires workshop participants to use an os with which they are often unfamiliar. for some partici- pants, unfamiliarity with linux can make the workshop more difficult than it would be if a more ubiquitous os was used. at demonstrations in a similar vein, members of the at team are often called upon to demonstrate the application at various profes- sional conferences and other venues. these demonstra- tions require the setup and population of a demonstration database with content for illustrating all of the applica- tion’s functions. one of the constraints posed by the demonstration scenario is the importance of using a local database instance rather than a networked instance, since network connections can be unreliable or outright unavailable (network connectivity being an issue we’ve all faced at conferences). another constraint is that portions of the demonstrations need some level of preparation (for example, knowing what search terms will return a non- empty result set), which must be customized for the unique content of a database. a final constraint is that, because portions of the demonstration (import and data merging) alter the state of the database, changes to the database must be easily reversible, or else new examples must be created before the database can be reused. building on our experience of using virtualization to implement multiple copies of an at installation, we evaluated the possibility of using the same technology for simplifying the setup necessary for demonstrating the at. as with the workshops, the use of a vm for at dem- onstrations allows for easy distribution of a prepopulated database, which can be used by multiple team members at disparate geographic locations and on different host oss. this significantly reduces the cost of creating (and recreating) demonstration databases. in addition, dem- onstration scripts can be shared between team members, creating additional time savings as well as facilitating team participation in the development and refinement of the demonstration. perhaps most important is the ability to roll back the vm to a specific state or snapshot of the database. this means the database can be quickly returned to its original state after being altered during a demonstration. overall, despite our initial anxiety about depending on the vm for presentations to large audi- ences, this solution has proven very useful, reliable, and cost-effective. at distribution implementing the at requires installing both the toolkit client and a database application such as mysql, instan- tiating an at database, and establishing the connection between database and client. for many potential cus- tomers of the at, the requirements for database creation and management can be a significant barrier due to inexperience with how such processes work and a lack of readily available it resources. many of these customers simply desire a plug-and-play version of the application that they can install and use without requiring technical assistance. it is possible to satisfy this need for a plug-and-play at by constructing a vm containing a fully installed and ready-to-use at application and database instance. this significantly reduces the number and difficulty of steps involved in setting up a functional at instance. the cus- tomer would only need to transfer the vm from a dvd or other source to their computer, download and install the vm reader, and then launch the at vm. they would then be able to begin using the at immediately. this removes the need for the user to perform database creation and management; arguably the most technically challenging portion of the setup process. users would still have the option of configuring the application (default values, lookup lists, etc.) in accord with the practices of their repository. batch processing catalog records the rapid growth of electronic resources is significantly changing the nature of library cataloging. not only are types of library materials changing and multiplying, the amount of e-resources being acquired increases each year. electronic book and music packages often contain tens of thousands of items, each requiring some level of catalog- ing. because of these challenges, staff are increasingly cataloging resources with specialized programs, scripts, and macros that allow for semiautomated record creation and editing. such tools make it possible to work on large sets of resources—work that would not be financially possible to perform manually item by item. however, the specialized configuration of the workstation required for using these automated procedures makes it very dif- ficult to use the workstation for other purposes at the same time. in fact, user interaction with the workstation while the process is running can cause a job to terminate prior to completion. in either scenario, productivity is compromised. virtualization offers an excellent remedy to this prob- lem. a virtual machine configured for semiautomated batch processing allows for unused resources on the workstation to process the batch requests in an isolated environment while, at the same time and on the same machine, the user is able to work on other tasks. in cases employing virtualization in library computing | hutt et al. where the user’s machine is not an ideal candidate for virtualization, the vm can be hosted via a hypervisor- based solution, and the user can access the vm with familiar remote access tools such as remote desktop in windows xp. secure sandbox in addition to challenges posed by increasingly large quantities of acquisitions, the ucsd libraries is also encountering an increasing variety of library material types. most notable is the variety and uniqueness of digital media acquired by the library, such as specialized programs to process and view research data sets, new media formats and viewers, and application installers. cataloging some of these materials requires that media be loaded and that applications be installed and run to inspect and validate content. but running or opening these materials, which are sometimes from unknown sources, poses a security risk to both the user’s worksta- tion and to the larger pool of library resources accessible via the network. many installers require a user to have administrative privileges, which can pose a threat to net- work security. the virtual machine allows for a user to have admin- istrative privileges within the vm, but not outside of the vm. the user can be provided with the privileges needed for installing and validating content without modifying their privileges on the host machine. in addition, the vm can be isolated by configuring its network connection so that any potential security risks are limited to the vm instance and do not extend to either the host machine or the network. laptop classroom instructors at the ucsd libraries need a laptop class- room that meets the usual requirements for this type of service (mobility, dependability, etc.) but also allows for the variety of computing environments and applica- tions in use throughout our several library locations. in a least-common-denominator scenario, computers are configured to meet a general standard (usually microsoft windows with a standard browser and office suite) and allow minimal customization. while this solution has its advantages and is easy to configure and maintain from the it perspective, it leaves much to be desired for an instructor who needs to use a variety of tools in the classroom, often on demand. the goal in this case is not to settle for a single generic build but instead look for a solution that accommodats three needs: n the ability to switch quickly between different customized os configurations n the ability to add and remove applications on demand in a classroom setting n the ability to restore a computer modified during class to its original state of course, regardless of the approach taken, the lap- tops still needed to retain a high level of system security, application stability, and regular hardware maintenance. after a thorough review of the different technologies and tools already in use in the libraries, we determined that virtualization might also serve to meet the require- ments of our laptop classroom. the need to support multiple users and multiple vms makes this scenario an ideal candidate for hypervisor-based virtualization. we decided to use vdi (virtual desktop infrastructure), a commercially available hypervisor product from vmware. vmware is one of the largest providers of virtualization software, and we were already familiar with several itera- tions of its host-based vm services. the core of our project plan consists of a base vm to be created and managed by our it department. to sup- port a wide variety of applications and instruction styles, instructors could create a customized vm specific to their library’s instruction needs with only nominal assistance from it staff. the custom vm would then be made avail- able on demand to the laptops from a central server (as depicted in figure above). in this manner, instructors could “own” and maintain a personal instructional com- puting environment, while the classroom manager could still ensure the laptop classroom as a whole maintained the necessary secure software environment required by it. as an added benefit, once these vms are established, they could be accessed and used in a variety of diverse locations. n considerations for implementation before implementing any virtualization solution, in-depth analysis and testing is needed to determine which type of solution, if any, is appropriate for a specific use case in a specific environment. this analysis should include three major areas of focus: user experience, application perfor- mance in the virtualized environment, and effect on the enterprise infrastructure. in this section of this paper, we review considerations that, in hindsight, we would have found to be extremely valuable in the ucsd libraries’ various implementations of virtualization. user experience traditionally, system engineers have developed systems and tuned performance according to engineering metrics (e.g., megabytes per second and network latency). while such metrics remain valuable to most assessments of a information technology and libraries | september computer application, performance assessments are being increasingly defined by usability and user experience fac- tors. in an academic computing environment, especially in areas such as library computer labs, these newer kinds of performance measures are important indicators of how effectively an application performs and, indirectly, of how well resources are being used. virtualization can be implemented in a way that allows library users to have access to both the virtual- ized and host oss or to multiple virtualized oss. since virtualization essentially creates layers within the work- station, multiple os layers (either host or virtualized) can cause the users to become confused as to which os they are interacting with at a given moment. in that kind of implementation, the user can lose his or her way among the host and guest oss as well as become disoriented by differing features of the virtualized oss. for example, the user may choose to save a file to the desktop, but may not be aware that the file will be saved to the desktop of the virtualized os and not the host os. external device sup- port can also be problematic for the end user, particularly with regard to common devices such as flash drives. the user needs to be aware of which operating system is in use, since it is usually the only one with which an external device is configured to work. authentication to a system is another example of how the relationship between the host and guest os can cause confusion. the introduction of a second os implicitly creates a second level of authentication and authoriza- tion that must be configured separately from that of the host os. user privileges may differ between the host and guest os for a particular vm configuration. for instance, a user might need to remember two logins or at least enter the same login credentials twice. these unexpected differences between the host and guest os produce nega- tive effects on a user’s experience. this can be a critical factor in a time-sensitive environment such as a computer lab, where the instructor needs to devote class time to teaching and not to preparing the computers for use and navigating students through applications. interface latency and responsiveness latency (meaning here the responsiveness or “sluggish- ness” of the software application or the os) in any inter- face can be a problem for usability. developers devote a significant amount of time to improving operating systems and application interfaces to specifically address this issue. however, users will often be unable to rec- ognize when an application is running a virtualized os and will thus expect virtualized applications to perform with the same responsiveness as applications that are not-virtualized. in our experience, some vm implementa- tions exhibit noticeable interface latency because of inher- ent limitations of the virtualization software. perhaps the most notable and restrictive limitation is the lack of advanced d video rendering capability. this is due to the lack of support for hardware-accelerated graphics, thus adding an extra layer of communication between the application and the video card and slowing down performance. in most hardware-accelerated d applica- tions (e.g., google earth pro or second life), this latency is such a problem that the application becomes unusable in a virtualized environment. recent developments have begun to address and, in some cases, overcome these limitations. in every virtualization solution there is overhead for the virtualization software to do its job and delegate resources. in our experience, this has been found to cause an approximately – percent performance penalty. most applications will run well with little or moder- ate changes to configuration when virtualized, but the overhead should not be overlooked or assumed to be inconsequential. it is also valuable to point out that the combination of applications in a vm, as well as vms running together on the same host, can create further performance issues. traditional bottlenecks the bottlenecks faced in traditional library comput- ing systems also remain in almost every virtualization implementation. general application performance is usu- ally limited by the specifications of one or more of the following components: processor, memory, storage, and network hardware. in most cases, assuming adequate hardware resources are available, performance issues can be easily addressed by reconfiguring the resources for the vm. for example, a vm whose application is memory- bound (i.e., performance is limited by the memory avail- able to the vm), can be resolved by adjusting the amount of memory allocated to the vm. a critical component of planning a successful virtual- ization deployment includes a thorough analysis of user workflow and the ways in which the vm will be utilized. although the types of user workflows may vary widely, analysis and testing serve to predict and possibly avoid potential bottlenecks in system performance. enterprise impact when assessing the effect virtualization will have on your library infrastructure, it is important to have an accurate understanding of the resources and capabilities that will form the foundation for the virtualized infrastructure. it is a misconception that it is necessary to purchase state- of-the-art hardware to implement virtualization. not only are organizations realizing how to utilize existing hardware better with virtualization for specific projects, they are discovering that the technology can be extended employing virtualization in library computing | hutt et al. to the rest of the organization and be successfully inte- grated into their it management practices. virtualization does, however, impose certain performance requirements for large-scale deployments that will be used in a / production environment. in such scenarios, organizations should first compare the level of performance offered by their current hardware resources with the performance of new hardware. the most compelling reasons to buy new servers include the economies of scale that can be obtained by running more vms on fewer, more robust servers, as well as the enhanced performance supplied by newer, more virtualization-aware hardware. in addi- tion, virtualization allows for resources to be used more efficiently, resulting in lower power consumption and cooling costs. also, the network is often one of the most overlooked factors when planning a virtualization project. while a local virtualized environment (i.e., a single computer) may not necessarily require a high performance network environment, any solution that calls for a hypervisor-based infrastructure requires considerable planning and scaling for bandwidth requirements. the current network hard- ware available in your infrastructure may not perform or scale adequately to meet the needs of this vm use. again, this highlights the importance of thorough user workflow analyses and testing prior to implementation. depending on the scope of your virtualization project, deployment in your library can potentially be expen- sive and can have many indirect costs. while the initial investment in hardware is relatively easy to calculate, other factors, such as ongoing staff training and system administration overhead, are much more difficult to determine. in addition, virtualization adds an additional layer to oftentimes already complex software licensing terms. to deal with the increased use of virtualization, software vendors are devoting increasing attention to the intricacies of licensing their products for use in such environments. while virtualization can ameliorate some licensing constraints (as noted in the at workshop use case), it can also conceal and promote licensing violations, such as multiple uses of a single-license applications or access to license-restricted materials. license review is a prudent and highly recommended component of implementing a virtualization solution. finally, concern- ing virtualization software itself, it also should be noted that while commercial vm companies usually provide plentiful resources for aiding implementation, several worthy open-source options also exist. as with any open- source software, the total cost of operation (e.g., the costs of development, maintenance, and support) needs to be considered. n conclusion as our use cases illustrate, there are numerous potential applications and benefits of virtualization technology in the library environment. while we have illustrated a number of these, many more possibilities exist, and fur- ther opportunities for its application will be discovered as virtualization technology matures and is adapted by a growing number of libraries. as with any technology, there are many factors that must be taken into account to evaluate if and when virtualization is the right tool for the job. in short, successful implementation of virtualization requires thoughtful planning. when so implemented, virtualization can provide libraries with cost-effective solutions to long-standing problems. references and notes . alessio gaspar et al., “the role of virtualization in com- puting education,” in proceedings of the th sigcse technical symposium on computer science education (new york: acm, ): – ; paul ghostine, “desktop virtualization: stream- lining the future of university it,” information today , no. ( ): ; robert p. goldberg, “formal requirements for virtu- alizable third generation architectures,” in communications of the acm , no. (new york: acm, ): – ; and karissa miller and mahmoud pegah, “virtualization: virtually at the desktop,” in proceedings of the th annual acm siguccs con- ference on user services (new york: acm, ): – . . for other, non–ucsd use cases of virtualization, see joel c. adams and w. d. laverell, “configuring a multi-course lab for system-level projects,” sigcse bulletin , no. ( ): – ; david collins, “using vmware and live cd’s to con- figure a secure, flexible, easy to manage computer lab envi- ronment,” journal of computing for small colleges , no. ( ): – ; rance d. necaise, “using vmware for dual operating systems,” journal of computing in small colleges , no. ( ): – ; and jason nieh and chris vaill, “experiences teaching operating systems using virtual platforms and linux,” sigcse bulletin , no ( ): – . . h. andrés lagar-cavilla, “vmgl (formerly xen-gl): opengl hardware d acceleration for virtual machines,” www .cs.toronto.edu/~andreslc/xen-gl/ (accessed oct. , ). editor’s note: we have an excellent editorial board for this journal and with this issue we’ve decided to begin a new column. in each issue of ital, one of our board members will reflect on some question related to technol- ogy and libraries. we hope you find this new feature thought-provoking. enjoy! any librarian who has been following the profes-sional literature at all in the past ten years knows that there has been an increasing emphasis on user-centeredness in the design and creation of library services. librarians are trying to understand and even anticipate the needs of users to a degree that’s perhaps unprecedented in the history of our profession. it’s no mystery as to why. we now live in a world where global computer networks link users directly with information in such a way that often, no middleman is required. users are exploring information on their own terms, at their own convenience, sometimes even using technolo- gies and systems that they themselves have designed or contributed to. at the same time, most libraries are feeling a finan- cial pinch. resources are tight, and local governments, institutions of higher education, and corporations are all scrutinizing their library operations more closely, ask- ing “what have you done for me lately?” the unspoken coda is “it better be something good, or i’m cutting your funding.” the increasing need to justify our existence, together with our desire to build more relevant services, is driving an increased interest in assessment. how do we know when we’ve built a successful service? how do we define “success?” and, perhaps most importantly, in a world filled with technologies that are “here today, gone tomorrow,” how do we decide which ones are appropri- ate to build into enduring and useful services? as a library technologist, it’s this last question that concerns me the most. i’m painfully aware of how quickly new technologies develop, mature, and fade silently into that good night with nary a trace. it’s like watching pro- tozoa under a microscope. which of these can serve as the foundation for real, useful services? it’s obvious to me that if i’m going to choose well, it’s vital that i place these services in context—and not my context, the user context. in order to do that, i need to understand the users. how do they do their work? what are they most concerned with? how do they think about the library in relation to the research process? how do they use technology as part of that process? how does that process fit into the larger context of the assignment? to answer questions like these, librarians often turn to basic marketing techniques such as the survey or the focus group. whether we are aware of it or not, the emphasis on user-centered design is making librarians into mar- keters. this is a new role for us, and one that most of us have not had the training to cope with. since most of us haven’t been exposed to marketing as a discipline of study, we don’t think of what we do as marketing, even when we use marketing techniques. but that’s what it is. so whether we know it or not, marketing, particularly market research, is important to us. marketing as a discipline is in the process of under- going some major changes right now. recent research in sociology, psychology, and neuroscience has uncovered some new and often startling insights into how human beings think and make decisions. marketers are strug- gling to incorporate these new models into their research methods, and to change their own thinking about how they discover what people want. i recently collided with this change when my own library decided to do a focus group to help us redesign our website. since we have a school of business, i asked one of our marketing profes- sors for help. her advice? don’t do it. as she put it: “you and the users would just be trading ignorances.” she then gave me a reading list, which included how customers think by gerald zaltman, which i now refer to as “the book that made marketing sexy.” zaltman’s book pulls together a lot of the recent research on how people think, make choices, and remem- ber. some of it is pretty mind-blowing: n % of human reasoning is unconscious. it hap- pens at a level we are barely aware of. n we think in images much more than we do in lan- guage n social context, emotion, and reason are all involved in the decision-making process. without emotion, we literally are unable to make choices. n all human beings use metaphors to explain and understand the world around them. metaphor is the bridge between the rational and emotional parts of the decision-making process. n memory is not a collection of immutable snapshots we carry around in our heads. it’s much more like a narrative or story—one that we change just by remembering it. our experience of the past and present are inextricably linked—one is constantly influencing the other. heady stuff. if you follow many of these ideas to their logical conclusions, you end up questioning the value of many traditional marketing techniques, such as surveys and focus groups. for example, if the social context in information technology and libraries | june kyle felker (felkerk@wlu.edu) is an ital editorial board member, – , and technology coordinator at washington and lee university library in lexington, virginia. editorial board thoughts kyle felker ital board member’s column | felker which a decision is made is important, then surveys are often going to yield false data, since the context in which the person is deciding to tick off this or that box is very different from the context in which they actually decide to use or not use your service or product. asking users “what services would be useful” in a focus group won’t be effec- tive because you are only interviewing the users’ rational thought process—it’s at least as important to find out how they feel about the service, your library, the task itself, and how they perceive other people’s feelings on the subject. zaltman proposes a number of very different market- ing techniques to get a more complete picture of user decision making: n use lengthy, one-on-one interviews. interviewing the unconscious is tricky and takes trust, it’s some- thing you can’t do in a traditional focus group set- ting. n use images. we think in images, and images are a richer field for bringing unconscious attitudes to the surface. n use metaphor. invite interviewees to describe their feelings and experiences in metaphor. explore the metaphors they come up with to more fully under- stand all the context. if this sounds more like therapy than marketing to you, then your initial reaction is pretty similar to mine. but the techniques follow logically from the research zaltman presents. how many of us have done user assessment and launched a new service, only to find a less than warm reception for it? how many of us have had users tell us they want something, only to see it go unused when it’s implemented? zaltman’s model offers potential explanations for why this happens, and meth- ods for avoiding it. lest you think this has nothing to do with technol- ogy, let me offer an example: library facebook/myspace profile pages. there’s been a lot of debate on how effec- tive and appropriate these are. it seems to me that we can’t gauge how receptive users are to this unless we understand how they feel about and think about those social spaces. this is exactly the sort of insight that new market- ing techniques purport to offer us. in fact, if the research is right, and there is a social and emotional component to every choice a person makes, then that applies to every choice a user makes with regard to the library, whether it’s the choice to ask a question at the reference desk, the choice to use the library website, or the choice to vote on a library bond issue. librarians are doing a lot of things we never imagined we’d ever need or want to do. web design. archival digi- tization. tagging. perhaps it’s also time to acknowledge that what we do has an important marketing component, and to think of ourselves as marketers (at least part time). i’m sold enough on zaltman’s ideas that i’m willing to try them out at my own institution, and i encourage you to do the same. reference . zaltman, gerald. how customers think: essential insights into the mind of the market (boston, mass.: harvard business school press, .) smartphones: a potential discovery tool | starkweather and stoward smartphones: a potential discovery tool wendy starkweather and eva stowers the anticipated wide adoption of smartphones by research- ers is viewed by the authors as a basis for developing mobile-based services. in response to the unlv libraries’ strategic plan’s focus on experimentation and outreach, the authors investigate the current and potential role of smart- phones as a valuable discovery tool for library users. w hen the dean of libraries announced a discovery mini-conference at the university of nevada las vegas libraries to be held in spring , we saw the opportunity to investigate the potential use of smartphones as a means of getting information and services to students. being enthusiastic users of apple’s iphone, we and the web technical support manager, developed a presentation highlighting the iphone’s poten- tial value in an academic library setting. because wendy is unlv libraries’ director of user services, she was interested in the applicability of smartphones as a tool for users to more easily discover the libraries’ resources and services. eva, as the health sciences librarian, was aware of a long tradition of pda use by medical professionals. indeed, first-year bachelor of science nursing students are required to purchase a pda bundled with select soft- ware. together we were drawn to the student-outreach possibilities inherent in new smartphone applications such as twitter, facebook, and myspace. n presentation our brief review of the news and literature about mobile phones in general provided some interesting findings and served as a backdrop for our presentation: n a total of percent of internet experts agreed that the mobile phone would be “the primary con- nection tool” for most people in the world by . the number of smartphone users is expected to top million by . there are currently million smartphone users, with sales in north america having grown percent in . n smartphones offer a combination of technologies, including gps tracking, digital cameras, and digi- tal music, as well as more than fifty-thousand spe- cialized apps for the iphone and new ones being designed for the blackberry and the palm pre. the palm pre offered less than twenty applications at its launch, but one million apllication downloads had been performed by june , , less than a month after launch. n the horizon report predicts that the time to adoption of these mobile devices in the educa- tional context will be “one year or less.” data gathered from campus users also was presented, providing another context. in march , a survey of university of california, davis (uc-davis) students showed that percent owned a smartphone. uc-davis is participating in apple’s university education forum. here at unlv, percent of students and percent of faculty and staff own a smartphone. the presentation itself highlighted the mobile appli- cations that were being developed in several libraries to enhance student research, provide library instruction, and promote library services. two examples were abilene christian university (http://www.acu.edu/technology/ mobilelearning/index.html), which in fall distrib- uted iphones and ipod touches to the incoming fresh- man class; and stanford university (http://www.stanford .edu/services/wirelessdevice/iphone/) which partici- pates in “itunes u” (http://itunes.stanford.edu/). if the libraries were to move forward with smartphone technol- ogies, it would be following the lead of such universities. readers also may be interested in joan lippincott’s recent concise summary of the implications of mobile technologies for academic libraries as well as the chap- ter on library mobile initiatives in the july library technology report. n goals: a balancing act ultimately the goal for many of these efforts is to be where the users are. this aspiration is spelled out in unlv libraries’ new strategic plan relating to infrastructure evolution, namely, “work towards an interface and system architecture that incorporates our resources, internal and external, and allows the user to access from their preferred starting point.” while such a goal is laudable and fits very well into the discovery emphasis of the mini-conference presentation, we are well aware of the need for further investigation before proceeding directly to full-scale devel- opment of a complete suite of mobile services for our users. of critical importance is ascertaining where our users are and determining whether they want us to be there and in what capacity. the value of this effort is demonstrated in booth’s research report on student interest in emerging technologies at ohio state university. the report includes the results of an extensive environmental survey of their wendy starkweather (wendy.starkweather@unlv.edu) is director, user services division, and eva stowers (eva.stowers @unlv.edu) is medical/health sciences librarian at the university of nevada las vegas libraries. information technology and libraries | december library users. the study is part of ohio state’s effort to actualize their culture of assessment and continuous learn- ing and to use “extant local knowledge of user popula- tions and library goals” to inform “homegrown studies to illuminate contextual nuance and character, customiza- tion that can be difficult to achieve when using externally developed survey instruments.” unlv libraries are attempting to balance early exper- imentation and more extensive data-driven decision-mak- ing. the recently adopted strategic plan includes specific directions associated with both efforts. for experimenta- tion, the direction states, “encourage staff to experiment with, explore, and share innovative and creative applica- tions of technology.” to that end, we have begun work- ing with our colleagues to introduce easy, small-scale efforts designed to test the waters of mobile technology use through small pilot projects. “text-a-librarian” has been added to our existing group of virtual reference service, and we introduced a “text the call number and record” service to our library’s opac in july . unlv libraries’ strategic plan helps foster the healthy balance by directing library staff to “emphasize data collec- tion and other evidence based approaches needed to assess efficiency and effectiveness of multiple modes and formats of access/ownership” and “collaborate to educate faculty and others regarding ways to incorporate library collections and services into education experiences for students.” action items associated with these directions will help the libraries learn and apply information specific to their users as the libraries further adopt and integrate mobile tech- nologies into their services. as we begin our planning in earnest, we look forward to our own set of valuable discoveries. references . janna anderson and lee rainie, the future of the internet iii, pew internet & american life project, http://www.pewinternet .org/~/media//files/reports/ /pip_futureinternet .pdf .pdf (accessed july , ). . sam churchill, “smartphone users: m by ,” blog entry, mar. , , dailywireless.org, http://www.daily wireless.org/ / / /smartphone-users- m-by- (accessed july , ). . mg siegler, “state of the iphone ecosystem: million devices and , apps,” blog entry, june , , tech crunch, http://www.techcrunch.com/ / / / -million-iphones -and-ipod-touches-and- -apps (accessed july , ). . jenna wortham, “palm app catalog hits a million down- loads,” blog entry, june , , new york times technology, http://bits.blogs.nytimes.com/ / / /palm-app-catalog- hits-a-million-downloads (accessed july , ). . larry johnson, alan levine, and rachel smith, horizon report, edition (austin, tex.: the new media consortium, ), http://www.nmc.org/pdf/ -horizon-report.pdf (accessed july , ). . university of california, davis. “more than % of cam- pus students own smartphones, yearly tech survey says,” technews, http://technews.ucdavis.edu/news .cfm?id= (accessed july , ). . university of nevada las vegas, office of informa- tion technology, “student technology survey report: – ,” http://oit.unlv.edu/sites/default/files/survey/survey results _students _ _ .pdf (accessed july , ). . joan lippincott, “mobile technologies, mobile users: implications for academic libraries,” arl bi-monthly report (dec. ), http://www.arl.org/bm~doc/arl-br- -mobile .pdf. (accessed july , ); ellyssa kroski, “library mobile initiatives,” library technology reports , no. (july ): – . . “unlv libraries strategic plan – ,” http://www .library.unlv.edu/about/strategic_plan - .pdf (accessed july , ): . . char booth, informing innovation: tracking student inter- est in emerging library technologies at ohio university (chicago: association of college and research libraries, ), http:// www.ala.org/ala/mgrps/divs/acrl/publications/digital/ ii-booth.pdf (accessed july , ); “unlv libraries strategic plan – ,” . . “unlv libraries strategic plan – ,” . . ibid. information technology and libraries | december andrew k. pacepresident’s message i n my first column, i mentioned that the lita board’s main objective is “to oversee the affairs of the division during the period between meetings.” of course, over- sight requires communication. sometimes this is among board members, or it’s an e-mail update, or a post to the lita-l discussion list, or even the articles in this journal. regardless, i see the cornerstone of “between-meeting oversight” as keeping the membership fully (or even partially) engaged from january through june and july through december. as a mea culpa for the board, but without placing the blame on any one individual, i am willing to concede that the board has not done an adequate job of engaging the membership between american library association (ala) meetings. while ala itself is addressing this problem with recommendations for virtual participation and online collaboration, lita should be at the forefront of setting the benchmark for virtual communication, par- ticipation, education, planning, and membership devel- opment. in an attempt to posit some solutions, as opposed to finding someone to blame, i first thought of the lita committees. which one should be responsible for commu- nicating lita opportunities and events to the member- ship using twenty-first-century technology? education? membership? web coordinating? program planning? publications? in the end, i was left with the choice of two evils: merge all the committees into one so that they can do everything or create a new committee to deal with the perceived problem. knowing that neither of those solutions will suffice, i’d like to put the onus back on the membership. maybe i’m trying to be a . librarian—crowdsourcing the prob- lem, that is, taking the task that might have been done by an individual or committee and asking for more of a community-driven solution. in the past, lita focused on the necessary technologies for crowdsourcing—discus- sion lists, blogs, and wikis—as if the technology alone could solve the problem. the bigwig taskforce and web coordinating committee have shouldered the burden of both implementing the technology and gaining philo- sophical consensus on its use—a daunting task that can easily appear chaotic. now that the technology is com- moditized (and generally embraced by ala at large and other divisions as well), perhaps it is time to embrace the philosophy of crowdsourcing. maybe it’s just because i have had cloud computing and web-scale architectures on the brain too much lately (having decided that it is impossible to serve two mas- ters—job and volunteer work—i shall forever endeavor to find the overlap between the two), but i sincerely believe that repeating the mantra that lita’s strength is its membership is not mere rhetorical lipservice. ebay is better for sellers because there are so many buyers; it is better for buyers because there are so many sellers. googledocs works for sharing documents better than a corporate wiki or microsoft sharepoint because it breaks down the barriers of domains, allowing the participants to determine who shares responsibility for producing something. barcamps are rising in popularity not only because of a content focus on open data, open source, and open access, but because of the participatory and user- generated style of the barcamp-style meetings. as a division of ala, lita has two challenges— leading the efforts of educating the membership, other divisions, and ala about impending sea changes in information technology, but also embracing these tech- nologies itself. we must eat our own dog food, as the say- ing goes. perhaps it is more fitting to suggest that lita must not only focus on getting technology to work, but putting technology to work. in the next few months, the lita board will be tack- ling lita’s strategic plan, which expires in . that means it is time not only to review the strategy—to edu- cate, to serve, to reach out—but also to assess the tactics employed to fulfill that strategy. you are probably reading this column in or after the month in which the strategic plan ends, which does not mean that we will be coasting into the ala midwinter meeting. on the contrary, i sin- cerely hope to gather enough information from commit- tees, task forces, members, and nonmembers in order for the lita leadership to come up with something strategi- cally meaningful going into the next decade. one year isn’t nearly long enough to see something this big through to completion. just as national politicians begin reelection campaigns as soon as they are elected, i suspect that ala divisional presidents begin think- ing about their legacy within the first couple months of office, if not before. but i hope, at least, to establish some groundwork, including a platform strategy that will allow the membership to maintain a connection with the board and with other members—to crowdsource solutions on a scale that has not been attempted in the past and that will solidify our future. and when we have a plan, you can trust that we will use all the available methods at our disposal to promote it and solicit your feedback. andrew k. pace (pacea@oclc.org) is lita president / and executive director, networked library services at oclc inc. in dublin, ohio. editorial | truitt marc truitteditorial marc truitt (marc.truitt@ualberta.ca) is associate director, bibliographic and information technology services, university of alberta libraries, edmonton, alberta, canada, and editor of ital. the catalog. love it? hate it? depending upon who is speaking, it may be cast as the ultimate portal that enables user access to all local and networked resources, or it may be a tool of byzantine complexity, comprehensible at best to but a small fraction of librarians able to navigate its bibliographic metadata encoded in an arcane s-era format. it is a rich trove of structured and controlled information assembled over decades by the work of countless dedicated catalogers and others. or, it is the now-obsolete product of a labor-intensive process of description and subject analysis that has no relevance in a web-centric world where “everything” is findable via the google search-box. its attempt to organize knowledge provides catalogers with a raison d’etre, but sends their colleagues and many users fleeing for simpler and more all-encompassing tools. it is our alpha and omega, our yin and our yang. few topics in librarianship—perhaps with the con- spicuous exception of that perennial library school favor- ite, our profession’s status as a profession—seem to provoke the range and depth of sentiment engendered by discussions of the place of the catalog. especially in recent years, criticism of the catalog has grown ever more strident, to the point where it has become commonplace in our profession’s literature to say that this most basic of library services “sucks.” as a consequence, librarians have increasingly fallen into one of two camps, with those critical of the catalog often simplistically characterized as favoring, and those defending it as opposing “change.” a number of initiatives have emerged in response to this ferment. some of these have focused on our bib- liographic metadata, and particularly on its ability to express the relationships and interconnectedness of the bibliographic universe. as we have traditionally cataloged whatever we had “in-hand,” our cataloging codes and encoding standards have done a very good job of manag- ing the description of bibliographic items; what they have not generally expressed well are the relationships among items. frbr and frad—the functional requirements for bibliographic records and the functional requirements for authority data—seem promising beginnings for addressing the relationship issues, although there are as-yet very few practical implementations. resource description and access (rda), the forthcoming successor to aacr , is designed around frbr concepts; it will be interesting to see how this plays out in the “real world;” equally interesting will be to what degree the present (or a modified) marc is able to express rda’s frbr-based relationship model. other approaches have focused on developing systems that are able to exploit our existing investment in biblio- graphic metadata in new and useful ways. the pioneering and best-known example of this, of course, is the discov- ery tool developed by a partnership of north carolina state university libraries and endeca, which premiered in early . this initiative included several innova- tive features not previously found in library catalogs, such as search result relevance ranking and the ability to perform faceted searching against a variety of controlled- vocabulary indices (subject/topical, form/genre, date, etc.) ncsu’s endeca discovery tool spawned an entirely new product segment for the catalog: major ils vendors have scrambled to develop their own next-gen products, combining relevancy and facets with additional function- ality such as web . social and collaborative tools and enhanced federated searching capabilities. the result of all this activity has been the first cross-platform growth opportunity for ils vendors since the development of resource-linking tools and the erm. we at ital have watched these trends with keen interest and have published works describing many of the major developments vis-a-vis the catalog in recent years. indeed, since late , ital has published at least eleven major papers on various topics related to improv- ing the catalog. with our publication of jennifer bowen’s report on the first phase outcomes of the university of rochester’s extensible catalog (xc) project in this issue of ital, we continue our commitment to publish important research in this area. the rochester project is noteworthy, both for its modular and metadata-focused approach and for its high visibility as an open source effort that has received significant support from the andrew w. mellon foundation. i predict that this paper will quickly take its place among the other ground-breaking works on the catalog that ital has published, and i’ll eagerly be await- ing the next progress report on the xc. n “must-reads” dept. okay, so i may not be the first out of the gate with this one, but for those of you who haven’t looked at it yet, trust me, you’ll want to. jonathan zittrain’s the future of the internet and how to stop it (yale university press, ), which divides the internet into “generative” tech- nologies such as the pc, and proprietary appliances such as the iphone, may or may not resonate with you, but i think it could well become the next big debate about where the net is and where it should be going. grab a copy and read it today. information technology and libraries | september success factors and strategic planning: rebuilding an academic library digitization program cory lampert and jason vaughan this paper discusses a dual approach of case study and research survey to investigate the complex factors in sustaining academic library digitization programs. the case study involves the background of the university of nevada, las vegas (unlv) libraries’ digitization pro- gram and elaborates on the authors’ efforts to gain staff support for this program. a related survey was admin- istered to all association of research libraries (arl) members, seeking to collect baseline data on their digital collections, understand their respective administrative frameworks, and to gather feedback on both negative obstacles and positive inputs affecting their success. results from the survey, combined with the authors’ local experience, point to several potential success fac- tors including staff skill sets, funding, and strategic planning. e stablishing a successful digitization program is a dialog and process already undertaken or cur- rently underway at many academic libraries. in , according to an institute of museum and library services report, “thirty-four percent of academic librar- ies reported digitization activities within the past months.” nineteen percent expect to be involved in digi- tization work in the next twelve months, and forty-four percent beyond twelve months. more current statistics from a subsequent study in reflected that digitiza- tion work has both continued and expanded, with half of all academic libraries performing digitization activi- ties. fifty-five percent of arl libraries responded to a survey informing part of the association of research libraries (arl) study managing digitization activities; of these, percent of the respondents indicated engagement in digitization. the ithaka study key stakeholders in the digital transformation in higher education found that nearly percent of large academic libraries either already have or plan to have digital repositories. with digitization becoming the norm in many institutions, the time is right to consider what factors contribute to the success and rapid growth of some library digitization programs while other institutions find digitization chal- lenging to sustain. the evolution of digitization at the unlv libraries is doubtless a journey many institutions have undertaken. over the past couple of years, those responsible for such a program at the unlv libraries have had the opportu- nity to revitalize the program and help collaboratively address some key philosophical questions that had not been systematically asked before, let alone answered. associated with this was a concerted focus to engage other less involved staff. one goal was to help educate them on academic digitization programs. another goal was to provide an opportunity for input on key ques- tions related to the programs’ strategic direction. as a subsequent action, the authors conducted a survey of other academic libraries to better understand what fac- tors have contributed to their programs’ own success as well as challenges that have proven problematic. many questions asked of our library staff in the planning and reorganization process were asked in the survey of other academic libraries. while the unlv libraries have undertaken what is felt are the proper structural steps and have begun to author policies and procedures geared toward an efficient operation, the authors wanted to bet- ter understand the experiences, key players, and underly- ing philosophies of other institutional libraries as theses pertain to their own digitization program. the following article provides a brief context relating the background of the unlv libraries’ digitization program and elaborates on the authors’ efforts toward educating library col- leagues and gaining staff buy-in for unlv’s digitization program—a process that countless other institutions have no doubt experienced, led, or suffered. the administered survey to arl members dealt with many topics similar to those that arose during the authors’ initial planning and later conversations with library staff, and as such, survey questions and responses are integrated in the following discussion. the authors administered a -question survey to the members of the arl. the focus of this survey was different from the previously mentioned arl study managing digitization activities, though several of the questions overlapped to some degree. in addition to demographic or concrete factual types of questions, the unlv libraries digitization survey had several ques- tions focused on perceptions—that is, staff support, administrative support, challenges, and benefits. areas of overlap with the earlier arl survey are mentioned in the appropriate context. though unlv isn’t a member of the arl, we consider ourselves a research library, and, regardless, it was a convenient way to provide some structure to the survey. survey responses were collected for a forty-five-day period from mid-june to late july, . through visiting each and every arl library’s web- site, the authors identified the individuals that appeared to be the “leaders” of the arl digitization programs, with instructions to forward the message to a colleague if cory lampert (cory.lampert@unlv.edu) is digitization projects librarian and jason vaughan (jason.vaughan@unlv.edu) is director, library technologies, university of nevada las vegas. success factors and strategic planning | lampert and vaughan they themselves had been incorrectly identified. this was very tricky, and revealed numerous program structures in place, differences between institutions in promoting their collections, and so on. the authors didn’t necessar- ily start with the presumption that all arl libraries even have a digitization program, but most (but not all) either seemed to have a formal organized digitization program with staffing, or at least had digitized and made available something, even if only a single collection. we e-mailed a survey announcement and a link to the survey to the targeted individuals, with a follow-up reminder a month later. responses were anonymous, and respondents were allowed to skip questions; thus the number of responses for the twenty-six questions making up the survey ranged from a low of thirty ( . percent) to a high of forty-four responses ( . percent). the average number of responses for each of the questions was . , yield- ing an overall response rate of . percent. questions were of three types: multiple choice (select one answer), multiple choice (mark all that apply), and open text. in addition, some of the multiple choice questions allowed additional open text comments. survey responses appear in appendix a. n context of the unlv libraries’ digitization program “digital collection,” for the purpose of the unlv library digitization survey, was defined as a collection of library or archival materials converted to machine-readable format to provide electronic access or for preservation purposes; typically, digital collections are library-created digital copies of original materials presented online and organized to be easily searched. they may offer features such as: full text search, brows- ing, zooming and panning, side by side comparison of objects, and export for presentation and reuse. one question the survey asked was “what year do you feel your library published its first ‘major’ digital collec- tion?” responses ranged from to ; the general average of all responses was . the earlier arl study found as the year most respondents began digiti- zation activities. mirroring this chronology, the unlv libraries has been active in designing digital projects and digitizing materials from library collections since the late s. technical web design expertise was developed in the cataloging unit (later renamed bibliographic and metadata services), and some of the initial efforts were to create online galleries and exhibits of visual materials from special collections, such as the jeanne russell janish ( ) exhibit. subsequently, the unlv libraries pur- chased the contentdm digital collection management software, providing both back-end infrastructure and front-end presentation for digital collections. later, the first digitization project with search functionality was created in partnership with special collections and was funded by a unlv planning initiative award received in . the early las vegas ( ) project focused on las vegas historical material and was designed to guide users to search, retrieve, and manipulate results using contentdm software to query a database. unlv’s development corresponds with regional developments in utah in , when “the largest academic institutions in utah were just beginning to develop digital imaging projects.” data from the imls study showed that, in the twelve months prior to the study release in , the majority of larger academic libraries had digitized between one and five hundred images for online presen- tation. in terms of staffing, digitization efforts occur in a wide variety of configurations, from large departments to solo librarians managing volunteers. for institutions with rec- ognized digitization staff, great variations exist between institutions in terms of where in the organizational chart digitization staff are placed. boock and vondacek’s research revealed that, of departments involved in digi- tization, special collections, archives, technical services, and newly created digital library units are where digiti- zation activities most commonly take place. a majority of respondents to the arl study indicated that some or all activities associated with digitization are distrib- uted across various units in the library. in , the unlv libraries created a formal department within the knowledge access management division—web and digitization services (wds)—initially comprising five staff focused on the development of the unlv libraries’ public website, the development of web-based applica- tions and databases to manage and efficiently present information resources, and the digitization and online presentation of library materials unique to the unlv libraries’ collections and of potential interest to a wider audience. augmenting their efforts were individuals in other departments helping with metadata standards, con- tent selection, and associated systems technical support. the unlv library digitization survey showed that the majority ( percent) of libraries that responded have at least one full-time staff member whose central job respon- sibility is to support digitization activities. this should not imply the existence of a fully staffed digitization program; the imls study found that . percent of larger academic libraries described themselves as lack- ing in sufficiently skilled technology staff to accomplish technology-related activities. central to any digitization program should be some structure in terms of how projects are proposed and subsequently prioritized. to help guide the priorities information technology and libraries | september of unlv’s infant wds department, a digital projects advisory committee was formed to help solicit and prioritize project ideas, and subsequently track the development of approved projects. this committee’s work could be judged as having mixed success partly because it met too infrequently, struggled with conflict- ing philosophical thoughts on digitization, and was confronted with the reality that staff that were needed to help bring approved ideas to fruition simply weren’t in place because of too many other library priorities draw- ing attention away from digitization. an evaluation of the lessons learned from these early years can be found in brad eden’s article. the unlv library digitization survey had several questions related to management and prioritization for digital projects and shows that despite the challenges of a committee-based decision- making structure, when a formal process is in place at all, . percent of survey respondents used a committee versus a single decision maker ( . percent) for deter- mining to whom projects are proposed for production. a follow-up question asked “how are approved projects ultimately prioritized?” the most popular response ( . percent) indicated “by a committee for review by multiple people,” followed by “no formal process” ( percent). “by a single decision maker” was selected by . percent of the respondents. the earlier arl study asked a somewhat related question: “who makes deci- sions about the allocation of staff support for digitiza- tion efforts? check all that apply.” out of seven possible responses, the three most popular were “head of cen- tralized unit,” “digitization team/committee/working group,” and “other person”; the other person was most often in an administrative capacity, such as a dean, director, or department head. administrative support for a program was another variable the unlv library digitization survey investi- gated. the survey asked respondents to rate, on a scale of one to five, “how would you characterize current support for digitization by your library’s administration?” more than percent of responses indicated “consistent sup- port,” followed by percent of respondents indicating “very strong support, top priority,” . percent ranking support as neutral, and . percent claiming “minimal support” or “very little support, or some resistance.” it was also clear from some of the other questions’ responses that the dean or director’s support (or lack thereof) can have dramatic effects on the digitization program. brought change to the unlv libraries in the form of a new dean. well-suited for the digitization program, she came from california, a state very heavily engaged and at the forefront of digitization within the library and larger aca- demic environment. one of her initiatives was a retooling of the digitization program at the unlv libraries, and her enthusiasm reflects a growing awareness of administrators regarding the benefits of digitization. n reorganization, library staff engagement, and decision making in , two new individuals joined unlv libraries’ web and digitization services department, the digitization projects librarian (filling a vacancy), and the web tech- nical support manager (a new position). a bit later, the systems department (providing technical support for the web and digitization servers, among other things), and the wds department were combined into a single unit and renamed library technologies. collectively, these changes brought new and engaged staff into the digitiza- tion program and combined under one division many of the individuals responsible for digital collection creation and support. perhaps more subtlety, this arrangement also provided formal acknowledgement of the impor- tance and desire of publishing digital collections. with the addition of new staff and a reorganization, a piece still missing was a resuscitation of library stake- holders to help solicit, prioritize, and manage the cre- ation of digital collections and an overall vision guiding the program. while the technical expertise, knowledge of metadata and imaging standards, and deep-rooted knowledge of digitization programs and concepts existed within the library technologies staff, other knowledge didn’t—primarily in-depth knowledge of the unlv libraries’ special collections and a track record of deep engagement with college faculty and the educational curriculum. similar to other organizations, the unlv libraries had not only created a new unit, but was also poised to introduce cross-departmental project groups that would collaborate on digitization activities. in their study of arl and greater western library association (gwla) libraries, book and vondracek found that this was the most commonly used organizational structure. knowledge of the concepts of a digitization program and what is involved in digitizing and sustaining a collec- tion was not widespread among other library colleagues. acknowledged, but not guaranteed up front for the unlv libraries, was the likely eventual reformation of a group of interested and engaged library stakeholders charged to solicit, prioritize, and provide oversight of the unlv libraries’ digitization program. for various reasons, the authors wanted to garner staff buy-in to the highest degree possible. apart from wanting less informed col- leagues to understand the benefits of a digitization pro- gram, it was also likely that such colleagues would help solicit projects through their liaison work with programs of study across campus. one unlv library digitization survey question asked, “how would you characterize support for digitization in your library by the majority of those providing content for digitization projects?” “consistent support” was indicated by . percent of respondents; . percent indicated “very strong support, top priority,” . percent indicated neutrality, and . success factors and strategic planning | lampert and vaughan percent indicated either minimal support or even some resistance. to help garner staff buy-in and set the stage for revitalizing the unlv libraries’ digitization efforts, we began laying the groundwork to educate and engage library staff in the benefits of a digitization program. this work included language successfully woven into the unlv libraries’ strategic plan and an authored white paper posing engaging questions to the larger library audience related to the strategic direction of the program. finally, we planned and executed two digitization work- shops for library staff. n the strategic plan one unlv library digitization survey question asked, “is the digitization program or digitization activities referenced in your library’s strategic plan?” a total of . percent indicated yes, with an additional percent indicating no specific references, but rather implied ref- erences. only . percent indicated that the digitization program was not referenced in any manner in the stra- tegic plan, while, surprisingly, responses ( . percent) indicated that their library doesn’t have a strategic plan. the unlv libraries’ strategic plan is an important docu- ment authored with wide feedback from library staff, and it exemplifies the participatory decision-making process in place in the library. the current iteration of the strategic plan covers – and includes various goals with supporting strategies and action items. in addition, all action items have associated assessment metrics and library staff responsible for championing the action items. departmental annual reports explicitly reference progress toward strategic plan goals. as such, if goals related to the digitization program appear in the strategic plan, that’s a clear indication, to some degree, of staff buy-in in acknowledging the significance of the digitization pro- gram. fortunately, digitization efforts figure prominently in several goals, strategies, and action items, including the following: n increasingly provide access to digital collections and services to support instruction, research, and outreach while improving access to the unlv libraries’ print and media collections. n provide greater access to digital collections while continuing to build and improve access to collec- tions in all formats to meet the research and teach- ing needs of the university. identify collections to digitize that are unique to unlv and that have a regional, national, and international research inter- est. create digital projects utilizing and linking col- lections. develop and adapt metadata and scanning standards that conform to national standards for all formats. provide content and metadata for regional and national digital projects. continue to develop expertise in the creation and management of digi- tal collections and information. collaborate with faculty, students, and others outside the library in developing and presenting digital collections. n be a comprehensive resource for the documenta- tion, investigation, and interpretation of the com- plex realities of the las vegas metropolitan area and provide an international focal point for the study of las vegas as a unique urban and cultural phenomenon. facilitate real and digital access to materials and information that document the his- torical, cultural, social, and environmental setting of las vegas and its region by identifying, collect- ing, preserving, and managing information and materials in all formats. identify unique collections that strengthen current collections of national and international significance in urban development and design, gaming, entertainment, and architec- ture. develop new access tools and enhance the use of current bibliographic and metadata utilities to provide access to physical and digital collections. develop web-based digital projects and exhibits based upon the collections. an associated capital campaign case statement associ- ated with the strategic plan lists several gift opportunities that would benefit various aspects of the unlv libraries; several of these include gift ideas related to the digitiza- tion of materials. n the white paper another important step in laying the groundwork for the digitization program was a comprehensive white paper authored by the recently hired digitization projects librar- ian. the finished paper was originally given to the dean of libraries and thereafter to the administrative cabinet, and eventually distributed to all library staff. the out- line of this white paper is provided as appendix b. the purpose of the white paper was multifaceted. after a brief historical context, the white paper addressed per- haps the single most important aspect of a digitization program—program planning—developing the strategic goals of the program, selecting and prioritizing projects though a formal decision-making process, and managing initiatives from idea to reality through efficient project teams. this first topic addressing the core values of the program had a strong educational purpose for the entire library staff—the ultimate audience of the paper. as part of its educational goal, the white paper enumerated the various strengths of digitization and why an institution information technology and libraries | september would want to sustain a digitization program (providing greater worldwide access to unique materials, promoting and supporting education and learning when integrated with the curriculum, etc.). it defined distinctions between an ephemeral digital exhibit and a long-term published and maintained collection. it discussed the various com- ponents of a digital collection—images, multimedia, metadata, indexing, thematic presentation (and the pref- erence to be unbiased), integration with other digital col- lections and the library website, etc. it posited important questions on sustenance and assessment, and defined concepts such as refreshing of data and migration of data to help set the stage for future philosophical discussions. given the myriad reasons one might want to publish a digital collection, checked by the reality that all the rea- sons and advantages may not be realized or given equal importance, the white paper listed several scenarios and asked if each scenario was a strong underlying goal for our program—in short, true or false: n “the libraries are interested in digitizing select unique items held in our collection and providing access to these items in new formats.” n “the libraries are interested in digitizing whole runs of an information resource for access in new formats.” n “the libraries should actively pursue funding to support major digitization initiatives.” n “the libraries should take advantage of the unique publicity, promotion, and marketing opportunities afforded by a digital project/program.” continuing with a purpose of defining boundaries of the new program, the paper asked questions related to audience, required skill sets, and resources. the second primary topic introduced the selection and prioritization of the items and ideas suggested for digiti- zation. it posed questions related to content criteria (why does this idea warrant consideration? would complex or unique metadata be required from a subject specialist?) and listed various potential evaluative measures of proj- ect ideas (should we do this if another library is already doing a very similar project?). technical criteria consider- ations were enumerated, touching on interoperability of collections in different formats, technical infrastructure considerations, and so on. multiple simultaneous ideas beg for prioritization, and the white paper proposed a formal review process and the library staff and skill sets that would help make such a process successful. the third primary topic focused on the details of carrying an approved idea to reality, and strengthened the educational purpose of the white paper. it described the general planning steps for an approved project and included a list of typical steps involved with most digital projects—scanning; creating metadata, indexes, and controlled vocabulary; coding and designing the web interface; loading records into unlv libraries’ contentdm system; publicizing the launch of the proj- ect; and assessing the project after completion. one unlv library digitization survey question was related to thir- teen such skills the unlv libraries identified as critical for a successful digitization program. the question asked respondents to rate skill levels possessed by personnel at their library, based on a five-point scale (from one to five: “no expertise,” “very limited expertise,” “working knowledge/enough to get by,” “advanced knowledge,” and “tremendous expertise”). neither “no expertise” nor “very limited expertise” garnered the highest number of responses for any of the skills. the overall rating average of all thirteen skills was . out of . the skills with the highest rating averages were “metadata creation/catalog- ing” . and “digital imaging/document scanning/post image processing/photography” with . . the skills with the lowest rating averages were “marketing and promotion” with . followed by “multimedia formats” with . . the unlv libraries’ white paper contained several appendixes that likely provided some of the richest content of the white paper. with the educational thrust completed, the appendixes drew a roadmap of “where do we want to go from here?” this roadmap suggested the revitalization of an overarching digital projects advisory committee, potential members of the committee, and functions of the committee. the committee would be responsible for soliciting and prioritizing ideas and track- ing the progress of approved ideas to publication. the appendixes also proposed project teams (which would exist for each project), likely members of the project teams, and the functions of the project team to complete day-to-day digitization activities. the liaison between the digital projects advisory committee and the project team would be the digitization projects librarian, who would always serve on both. the last page of the white paper provided an illustration highlighting the various steps proposed in the lifecycle of a digital project—from concept to reality. n digitization workshops several months after the white paper had been shared, the next step in restructuring the program and building momentum was sponsoring two forums on digitization. the first one occurred in november and included two speakers brought in for the event, roy tennant (formerly user services architect with the california digital library and now with oclc) and ann lally (head of the digital initiatives program at the university of washington libraries). this session consisted of a success factors and strategic planning | lampert and vaughan two-hour presentation and q&a to which all library staff were invited, followed by two breakout sessions. all three sessions were moderated by the digitization projects librarian. questions from these sessions are pro- vided in appendix c. the breakout sessions were each targeted to specific departments in the unlv libraries. the first focused on providing access to digital collec- tions (definitions of digital libraries, standards, designing useful metadata, accessibility and interoperability, etc.). the second focused on components of a well-built digital library (goals of a digitization program, content selection criteria, collaboration, evaluation and assessment, etc.). colleagues from other libraries in nevada were invited, and the forum was well attended and highly praised. the sessions were recorded and later made available on dvd for library staff unable to attend. this initial forum accomplished two important goals. first, it was an all- staff meeting offering a chance to meet, explore ideas, and learn from two well-known experts in the field. second, it offered a more intimate chance to talk about the technical and philosophical aspects of a digitization program for those individuals in the unlv libraries associated with such tasks. as a momentum-building opportunity for the digitization program, the forum was successful. the second workshop occurred in april . to gain initial feedback on several digitization questions and to help focus this second workshop, we sent out a survey to several dozen library staff—those that would likely play some role at some point in the digitization program. the survey contained questions focused on several the- matic areas: defining digital libraries, boundaries to the digitization program, users and audience, digital project design, and potential projects and ideas. it contained thirteen questions consisting of open-ended response questions, questions where the respondent ranked items on a five-point scale, and “select all that apply”–type questions. we distributed the survey to invitees to the second workshop, approximately three dozen individu- als; of those, eighteen (about percent) responded to most of the questions. the survey was closely tied to the white paper and meant to gauge early opinions on some of the questions posed by that paper. whereas the first workshop included some open q&a, the second ses- sion was structured as a hands-on workshop to answer some of the digitization questions and to illustrate the complexity of prioritizing projects. the second workshop began with a status update on the retooling of the unlv libraries’ digitization program. this was followed by an educational component that focused on a diagram that detailed the workflow of a typical digitization project and who was involved and that emphasized the fact that there is a lot of planning and effort needed to bring an idea to reality. in addition, we discussed project types and how digital projects can vary widely in scope, con- tent, and purpose. finally, we shared general results from the aforementioned survey to help set the stage for the structured hands-on exercises. the outline for this second workshop is provided in appendix d. one question of the unlv library digitization survey asked, “on a scale of to , how important are each of the factors in weighing whether to proceed with a proposal for a new digital collection project, or enhancement of an existing project?” eight factors were listed, and the five- point scale was used (from one to five: “not important,” “less important,” “neutral,” “important,” and “vitally important”). the average rating for all eight factors was . . the two most important factors were “collection includes unique items” ( . average rating) and “col- lection includes items for which there is a preservation concern or to make fragile items more accessible to the public” ( . average rating). the factors with the lowest average ratings were “collection includes integration of various media into a themed presentation” ( . average rating) followed by “collection involves a whole run of an information resource (i.e., such as an entire manuscript, newspaper run, etc.” ( . average rating). the earlier arl survey asked a somewhat related question, “what is/has been the purpose of these digitization efforts? check all that apply.” of the six possible responses (which differed somewhat from those in the unlv library digitization survey), the most frequent responses were “improved access to library collections,” “support for research,” and “preservation.” the earlier survey also asked the question, “what are the criteria for select- ing material to be digitized? check all that apply.” the most frequent responses were “subject matter,” “mate- rial is part of a collection being digitized,” and “rarity or uniqueness of the item(s).” the first exercise of the second digitization workshop focused on digital collection brainstorming. the authors provided a list of ten project examples and asked each of the six tables (with four colleagues each) to prioritize the ideas. afterward, a speaker from each table presented the prioritizations and defended their rankings. this exercise successfully illustrated to peers in attendance that different groups of people have different ideas about what’s important and what constitutes prime materials for digitization. the rankings from the varying tables were quite divergent. a related question asked of the arl libraries in the unlv library digitization survey was “from where have ideas originated for existing, published digital collection at your library?” and offered six choices. respondents could mark multiple items. the most chosen answer ( . percent) was “special collections, archives, or library with a specialized collection or focus.” the least chosen answer ( . percent) was “an external donor, friend of the library, community user, etc.” for the second part of the workshop exercise, each table came up with their own digital collection ideas, defined the audience and content of the proposal, and defended and information technology and libraries | september explained why they thought these were good proposals. fourteen unique and varied ideas were proposed, most of which were tightly focused on las vegas and nevada, such as “history of las vegas,” “unlv yearbooks,” “las vegas gambling and gamblers,” and “african american entertainers in las vegas.” other proposals were less tied to the area, such as a “botany collection,” “movie posters,” “children’s literature,” “architecture,” and “federal land management.” this exercise successfully showed that ideas for digital collections stretch across a broad spectrum, as broad as the individual brainchilden themselves. finally, in the last digitization workshop exercise, each table came up with specialties, roles, and skills of candidates who could potentially serve on the proposed committee, and defended their rationale—in other words, committee success factors. this exercise generated nine- teen skills seen as beneficial by one or more of the group tables. at the end of the workshop, we asked if others had alternate ideas to the proposed committee. none sur- faced, and the audience thought such a committee should be reestablished. this second workshop concluded with a brief discussion on next steps—drafting a charge for the committee, choosing members, and a plug for the expec- tation of subject liaisons working with their respective areas to help better identify opportunities for collabora- tion on digital projects across campus. n toward the future digital projects currently maintained by the unlv libraries include both static web exhibits in the tra- dition of unlv’s first digitization efforts, as well as several searchable contentdm–powered collections. the unlv libraries have also sought to continue col- laborative efforts, participating as project partners for the western waters digital library (phase ) and continu- ing in a regional collaboration as a hosting partner in the mountain west digital library. partnerships were shown in the unlv library digitization survey to garner increased buy-in for projects, with one respon- dent commenting that faculty partnerships had been “the biggest factor for success of a digital library proj- ect.” institutional priorities at unlv libraries reflect another respondent’s comment regarding “interesting archival collections” as a success factor. one recently launched unlv collection is the showgirls collection ( ), focused on a themed collection of historical mate- rial about las vegas entertainment history. another recently launched collection, the nevada test site oral history project ( ), recounts the memories of those affiliated with and affected by the nevada test site dur- ing the era of cold war nuclear testing and includes searchable transcripts, selected audio and video clips, and scanned photographs and images. with general library approval, the restructured digitization projects advisory committee was estab- lished in july with six members drawn from library technologies, special collections, the subject special- ists, and at large. the advisory committee has drafted and gained approval for several key documents to help govern the committee’s future work. this includes a col- lection development policy for digitization projects and a project proposal form to be completed by the individual or group proposing an idea for a digital collection. at the time of writing, the committee is just now at the point of advertising the project proposal form and process, and time will tell how successful these documents prove. in the unlv library digitization survey, . percent responded that a digitization mission statement or collec- tion development policy was in place at their institution. one goal at unlv is to “ramp up” the number of simul- taneous digitization projects underway at any one time at unlv. many items in the special collections are ripe for digitization. many of these are uncataloged, and digitiz- ing such collections would help promote these hidden treasures. related to ramping up production, one unlv library digitization survey question asked, “on average over the past three years, approximately how many new digital collections are published each year?” responses ranged from zero new collections to sixty. the average number of new collections added each year was . for the respondents who gave exact numerical answers. while this is perhaps double the unlv libraries’ current rate of production, it illustrates that increasing produc- tion is an achievable goal. staffing and funding for the unlv libraries’ digitiza- tion program have both seen increases over the past several years. a new application developer was hired, and a new graphics/multimedia specialist filled an existing vacancy. together, these staff have helped with projects such as modifying contentdm templates, graphic design, and multimedia creation related to digital projects, in addition to working on other web-based projects not necessarily related to the digitization program. another position has a job focus shifted toward usability for all things web- based, including digitization projects. in terms of funding, the two most recent projects at the unlv libraries are both the result of successful grants. the recently launched nevada test site oral history project was the result of two grants from the u.s. departments of education and energy. subsequently, a $ , lsta grant proposal seek- ing to digitize key items related to the history of southern nevada from to was funded for – , with the resulting digital collection publicly launched in may . this collection, southern nevada: the boomtown years, contains more than , items from several institutions, focused on the heyday of mining town life in southern success factors and strategic planning | lampert and vaughan nevada during the early twentieth century. this grant funded four temporary positions: a metadata specialist, an archivist, a digital projects intern, and an education consultant to help tie the digitized collection into the k– curriculum. grants will likely play a large role in the unlv libraries’ future digitization activities. the unlv library digitization survey asked, “has your institu- tion been the recipient of a grant or gift whose primary focus was to help efforts geared toward digitization of a particular collection or to support the overall efforts of the digitization program?” the question sought to determine if grants had played a role, and if so, whether it was primarily large grants (defined as > $ , ), small grants (< $ , ), or both. the major- ity of responses ( . percent), indicated a combination of both small and large grants had been received in sup- port of a project or the program. an additional . per- cent indicated that large grants had played a role, and . percent indicated that one or more small grants had played a role. two respondents ( . percent) indicated that no grants had been received or that they had not applied for any grants. the earlier arl survey asked the question, “what was/is the source of the funds for digitization activities? check all that apply.” of seven possible responses, “grant” was the second most fre- quent response, trailing only “library.” with an eye toward the future, the survey adminis- tered to arl libraries asked two blunt questions sum- marizing the overall thrust of the survey. one of the final open-ended survey questions asked, “what are some of the factors that you feel have contributed to the success of your institution’s digitization program?” forty respondents offered answers that ranged from list- ing one item to multiple items. several responses along the same general theme seemed to surface, which could be organized into rough clusters. in general, support from library administration was mentioned by a dozen respondents, with such statements as “consistent inter- est on the part of higher level administration,” “having support for the digitization program at an administra- tive level from the very beginning,” “good support from the library administration,” “support of the dean,” and, mentioned multiple times in the same precise language, “support from library administration.” faculty collabo- ration and interest across campus was mentioned by ten respondents, evidenced by statements such as “strong collaboration with faculty partners,” “support of faculty and other partners,” “interest from faculty,” “heavily involving faculty in particular . . . ensures that we can have continued funding since the faculty can lobby the provost’s office,” and “grant writing partnerships with faculty.” passionate individuals involved with the pro- gram and/or support from other staff in the libraries were mentioned by ten respondents, with comments such as “program management is motivated to achieve success,” “a strong department head,” “individual staff member ’s dedication to a project,” “commitment of the people involved,” “team work, different departments and staff willing to work together,” and “supportive individuals within the library.” having “good” content to digitize was mentioned by seven respondents, with statements such as “good content,” “collection strength,” “good collections,” and “availability of unique source materials.” strategic plan or goals integration was men- tioned in several responses, such as “strong financial commitment from the strategic plan” and “mainstream- ing the work of digital collection building into the stra- tegic goals of many library departments.” successful grants and donor cultivation were mentioned by four respondents. other responses were more unique, such as one respondent’s one-word response—“luck”—and other responses such as “nimbleness, willingness, and creativity,” and “a vision for large-scale production, and an ability to achieve it.” the final unlv library digitization survey question asked, “what are the biggest challenges for your institu- tion’s digitization program?” thirty-nine respondents provided feedback, and again, several variations on a theme emerged. the most common response, unsurpris- ingly, “not enough staffing,” was mentioned by eighteen respondents, with responses such as “lack of support for staffing at all necessary levels,” “the real problem is people, we don’t have enough staff,” “limited by staff,” and “we need more full-time people.” following this was (a likely related response) “funding,” mentioned by another nine respondents, with statements such as “funding for external digitization,” “identifying enough funding to support conversion,” “we could always use more money,” and, succinctly, “money.” related to staff- ing, specifically, six responses focused on technical staff or support from technical staff, such as “need more it (information technology) staff,” “need support from existing it staff,” “not enough application development staff,” and “limited technical expertise.” prioritization and demand issues surfaced in six responses, with responses such as “prioritizing efforts now that many more requests for digital projects have been submit- ted,” “prioritization,” “can’t keep up with demand,” and “everyone wants to digitize everything.” workflow was mentioned in four responses, such as “workflow bottlenecks,” “we need to simplify the process of getting materials into the repository,” and “it takes far longer to describe an object than to digitize it, thus creating bottlenecks.” “not enough space” was mentioned by three respondents, and “maintaining general library- wide staff support for the program” was mentioned by two respondents. the unlv libraries will keep in mind the experiences of our colleagues, as few, if any, libraries are likely immune to similar issues. information technology and libraries | september n conclusions the unlv library digitization survey revealed, not sur- prisingly, that not all libraries, even those of high stature, are created equally. many have struggled to some extent in growing and sustaining their digitization programs. many have numerous published projects, others have few or perhaps even none. administrative and fellow colleague support varies, as does funding. additional questions remain to be tackled at the unlv libraries. how precisely will we define success for the digitization program? by the number of published collections? by the number of successful grants executed? by the num- ber of image views or metadata record accesses? by the frequency of press in publications and word-of-mouth praise from fellow colleagues? ideas abound, but no definitive answers exist as of yet. at the larger level, other questions are looming. as libraries continue to promote themselves as relevant in the digital age, and promote themselves as a (or the) central partner in student learn- ing, to what degree will libraries’ digital collections be tied into the educational curriculum, whether at their own affiliated institutions or with k– in their own states as well as beyond? clearly the profession is chang- ing, with library schools creating courses and certificate programs in digitization. discussions about the integra- tion of various information silos, metadata crosswalk- ing, and item exposure in other online systems used by students will continue. library digitized collections are primary resources involved in such discussions. while these questions persist, it’s hoped that at a minimum, the unlv libraries have established the foundational struc- ture to foster what we hope will be a successful digitiza- tion program. references . institute for museum and library services, “status of technology and digitization in the nation’s museums and librar- ies report,” may , , www.imls.gov/publications/ techdig / report.pdf (accessed mar. , ). . institute for museum and library services, “status of technology and digitization in the nation’s museums and libraries report,” jan. , www.imls.gov/resources/ techdig /technology% bdigitization.pdf (accessed mar. , ). . rebecca mugridge, managing digitization activities, spec kit (washington, d.c.: association of research libraries, ): . . ross housewright and roger schonfeld, “ithaka’s studies of key stakeholders in the digital transformation in higher education,” aug. , , www.ithaka.org/research/ ithakas% % studies% of% key% stakeholders% in% the% digital% transformation% in% higher% education.pdf (accessed mar , ). . ibid. . university of nevada, las vegas university libraries, “jeanne russell janish, botanical illustrator: landscapes of china and the southwest,” oct. , , http://library.unlv .edu/speccol/janish/index.html (accessed mar. , ). . university of nevada, las vegas university librar- ies, “early las vegas,” http://digital.library.unlv.edu/early_ las_vegas/earlylasvegas/earlylasvegas.html (accessed mar. , ). . arlitsch, kenning, and jeff jonsson, “aggregating distrib- uted digital collections in the mountain west digital library with the contentdm multi-site server,” library hi tech , no. ( ): . . institute for museum and library services, “status of technology and digitization in the nation’s museums and libraries report.” . michael boock and ruth vondracek, “organizing for digitization: a survey,” portal: libraries and the academy , no. ( ), http://muse.jhu.edu/journals/portal_libraries_and_ the_academy/v / . boock.pdf (accessed mar. , ). . mugridge, managing digitization activities, . . institute for museum and library services, “status of technology and digitization in the nation’s museums and libraries report.” . brad eden, “managing and directing a digital proj- ect,” online information review , no. ( ), www.emerald insight.com/insight/viewpdf.jsp?contenttype=article& filename=html/output/published/emeraldfulltextarticle/ pdf/ .pdf (accessed mar. , ). . mugridge, managing digitization activities, – . . boock and vondracek, “organizing for digitization: a survey.” . university of nevada, las vegas university librar- ies, “university libraries strategic goals and objectives,” june , , www.library.unlv.edu/about/strategic_goals.pdf (accessed mar. , ). . mugridge, managing digitization activities, . . ibid, . . university of nevada, las vegas university librar- ies, “showgirls,” http://digital.library.unlv.edu/showgirls/ (accessed mar. , ). . university of nevada, las vegas university libraries, “nevada test site oral history project,” http://digital.library .unlv.edu/ntsohp/ (accessed mar. , ). . university of nevada, las vegas university librar- ies, “southern nevada: the boomtown years,” http://digital .library.unlv.edu/boomtown/ (accessed may , ). . mugridge, managing digitization activities, . success factors and strategic planning | lampert and vaughan appendix a. unlv library digitization survey responses . is the digitization program or digitization activities referenced in your library’s strategic plan? answer options ( responses total) response percent response count yes . no . not specifically, but implied . our library doesn’t have a strategic plan . . how would you characterize current support for digitization by your library’s administration? answer options ( responses total) response percent response count very strong support, top priority . consistently supportive . neutral . minimal support, . very little support, or some resistance . . how would you characterize support for digitization in your library by the majority of those providing content for digitization projects (i.e., regardless of whether those providing content have as a primary or a minor responsibility provisioning content for digitization projects)? answer options ( responses total) response percent response count very strong support, top priority . consistently supportive . neutral . minimal support . very little support, or some resistance . information technology and libraries | september . what year do you feel your library published its first “major” digital collection? major is defined as this was the first project deemed as having permanence and which would be sustained; it has associated metadata, etc. if you do not know, you may estimate or type “unknown.” responses ranged from to . . to date, approximately how many digital collections has your library published? (please do not include ephemeral exhibits that may have existed in the past but no longer are present or sustained.) responses ranged from to , s. the great majority of responses were under ; four responses were between and , and one response was “ , s.” success factors and strategic planning | lampert and vaughan . on average over the past years, approximately how many new digital collections are published each year? all but two responses ranged from to . one response was , one was . . what hosting platform(s) do you use for your digital collections (e.g., contentdm, etc.)? . does your institution have an institutional repository (e.g., dspace)? answer options ( responses total) response percent response count yes . no . . if the answer was “yes” in question , is your institutional repository using the same software as your digital collections? answer options ( responses total) response percent response count yes . no . information technology and libraries | september . is there an individual at your library whose central job responsibility is the development, oversight, and manage- ment of the library’s digitization program? (for purposes of this survey, central job responsibility means that percent or more of the employee’s time is dedicated to digitization activities.) answer options ( responses total) response percent response count yes . no . . are there regular, full-time staff at your library who have as their primary or one of their primary job responsi- bilities support of the digitization program? for this question, a primary job responsibility means that at least percent of their normal time is spent on activities directly related to supporting the digitization program or devel- opment of a digital collection. (mark all that apply) answer options ( responses total) response percent response count digital imaging/document scanning, post-image processing, photography . metadata creation/cataloging . archival research of documents included in a collection(s) . administration of the hosting server . grant writing/donor cultivation/program or collection marketing . project management . multimedia formats . database design and data manipulation . maintenance, customization, and/or configuration of digital asset management software or features within that software (e.g., contentdm) . programming languages . web design and development . usability . marketing and promotion . none of the above . . approximately how many individuals not on the full-time library staff payroll (i.e., student workers, interns, field- workers, volunteers) are currently working on digitization projects? answers ranged from to “approximately .” the majority of responses ( ) fell between and workers; twelve responses indicated more than ; several responses indicated “unknown.” success factors and strategic planning | lampert and vaughan . has your library funded staff development, training, or conference opportunities that directly relate to your digi- tization program and activities for one or more library staff members? answer options ( responses total) response percent response count yes, frequently, one or more staff have been funded by library administration for such activities . yes, occasionally, one or more staff have been funded by library administration for such activities . no, to the best of my knowledge, no library staff member has been funded for such activities . . where does the majority of digitization work take place? answer options ( responses total) response percent response count centralized in the library (majority of content digitized using library staff and equipment in one department) . decentralized (majority of content digitized in multiple library departments or outside the library by other university entities) . through vendors or outsourcing . hybrid of approaches depending on project . . on a scale of to ( being least important and being vitally important), how important are each of the factors in weighing whether to proceed with a proposal for a new digital collection project or enhancement of an existing project? answer options ( responses total) not important less important neutral important vitally important rating average response count collection includes item(s) for which there is a preservation concern or to make fragile item(s) more accessible to the public . collection includes unique items . collection involves a whole run of an information resource (e.g., an entire manuscript, newspaper run, etc.) . information technology and libraries | september answer options ( responses total) not important less important neutral important vitally important rating average response count collection includes the integration of various media (i.e., images, documents, audio) into a themed presentation . collection has a direct tie to educational programs and initiatives (e.g., university courses, statewide education programs, or k– education) . collection supports scholarly communication and/or management of institutional content . collection involves a collaboration with university colleagues . collection involves a collaboration with entities external to the university (e.g., public libraries, historical societies, museums) . . from where have ideas originated for existing, published digital collections at your library? in other words, have one or more digital collections been the brainchild of one of the following? (mark all that apply) answer options ( responses total) response percent response count library subject liaison or staff working with teaching faculty on a regular basis . library administration . special collections, archives, or library with a specialized collection or focus . digitization program manager . university staff or faculty member outside the library . an external donor, friend of the library, community user, etc. . (continued from previous page) success factors and strategic planning | lampert and vaughan . to whom are new projects first proposed to be evaluated for digitization consideration? answer options ( responses total) response percent response count to an individual decision-maker . to a committee for review by multiple people . no formal process . . how are approved projects ultimately prioritized? answer options ( responses total) response percent response count by a single decision-maker . by a committee for review by multiple people . by departments or groups outside of the library . no formal process . . are digitization program mission statements, selection criteria, or specific prioritization procedures in use? answer options ( responses total) response percent response count yes, one or more of these forms of documentation exist detailing process . yes, some criteria are used but no formal documentation exists . no documented process in use . . what general evaluation criteria do you employ to measure how successful a typical digital project is? (mark all that apply) answer options ( responses total) response percent response count log analysis showing utilization/record views of digital collection items . analysis of feedback or survey responses associated with the digital collection . publicity generated by, or citations referencing, digital collection . e-commerce sales or reproduction requests for digital images . we have no specific evaluation measures in use . information technology and libraries | september . has your institution been the recipient of a grant or gift whose primary focus was to help efforts geared toward digitization of a particular collection or to support the overall efforts of the digitization program? answer options ( responses total) response percent response count we have received one or more smaller grants or donations (each of which was $ , or less) to support a digital collection/program . we have received one or more larger grants or donations (each of which was greater than $ , ) to support a digital collection/program . we have received a mix of small and large grants or donations to support a digital collection/program . we have been unsuccessful in receiving grants or have not applied for any grants—grants and/or donations have not played any role whatsoever in supporting a digital collection or our digitization program . . how would you rate the overall level of buy-in for collaborative digitization projects between the library and external partners (an external partner is someone not on the full-time library staff payroll, such as other university colleagues, colleagues from other universities, etc.)? answer options ( responses total) response percent response count excellent . good . neutral . minimal . low or none . not applicable—our library has not yet published or attempted to publish a collaborative digital project involving individuals outside the library . . when considering the content available for digitization, which of the following statements apply? (mark all that apply) answer options ( responses total) response percent response count at my institution, there is a lack of suitable library collections for digitization . content providers regularly contact the digitization program with project ideas . the main source of content for new digitization projects comes from special collections, archives, other libraries with specialized collections (maps, music, etc.), or local cultural organizations (historical societies, museums) . success factors and strategic planning | lampert and vaughan answer options ( responses total) response percent response count the main source of content for new digitization projects comes from born digital materials (such as dissertations, learning objects, or faculty research materials) . content digitization is mainly limited by available resources (lack of staffing, space, equipment, expertise) . obtaining good content for digitization can be challenging . . various types of expertise are important in collaborative digitization projects. please rate the level of your local library staff’s expertise in the following areas ( – scale, with having no expertise and having tremendous expertise). answer options ( responses total) no expertise very limited expertise working knowledge/ enough to “get by” advanced knowledge tremendous expertise n/a rating average response count digital imaging/ document scanning, post image processing, photography . metadata creation/ cataloging . archival research of documents included in a collection . administration of the hosting server . grant writing/ donor cultivation . project management . multimedia formats . database design and data manipulation . (continued from previous page) information technology and libraries | september answer options ( responses total) no expertise very limited expertise working knowledge/ enough to “get by” advanced knowledge tremendous expertise n/a rating average response count digital asset management software (e.g., contentdm) . programming languages . web design and development . usability . marketing and promotion . . what are some of the factors that you feel have contributed to the success of your institution’s digitization program? survey responses were quite diverse because respondents were speaking to their own perceptions and institutional expe- rience. the general trend of responses are discussed in the body of the paper. . what are the biggest challenges for your institution’s digitization program? survey responses were quite diverse because respondents were speaking to their own perceptions and institutional expe- rience. the general trend of responses are discussed in the body of the paper. appendix b. white paper organization i. introduction ii. current status of digitization projects at the unlv libraries iii. topic : program planning a. are there boundaries to the libraries digitization program? what should the program support? b. what resources are needed to realize program goals? c. who is the user or audience? d. when selecting and designing future projects, how can high-quality information be presented in online for- mats incorporating new features while remaining un-biased and accurate in service provision? e. to what degree do digitization initiatives need their own identity versus heavily integrating with the libraries’ other online components, such as the general website? f. how do the libraries plan on sustaining and evaluating digital collections over time? g. what type of authority will review projects at completion? how will the project be evaluated and promoted? iv. topic : initiative selection and prioritization a. project selection: what content criteria should projects fall within in order to be considered for digitization and what is the justification for conversion of the proposed materials? (continued from previous page) success factors and strategic planning | lampert and vaughan b. project selection: what technical criteria should projects fall within in order to be considered for digitization? c. project selection: how does the project relate to, interact with, or complement other published projects and collections available globally, nationally, and locally? d. project selection and prioritization: after a project meets all selection criteria, resources may need to be eval- uated before the proposal reaches final approval. what information needs to be discussed in order to finalize the selection process, select between qualified project candidates, and begin the prioritization process for approved proposals? e. project prioritization: should we develop a formal review process? v. topic : project planning a. what are the planning steps that each project requires? b. who will be responsible for the different steps in the project plan and department workload? c. how can the libraries provide rich metadata and useful access points? d. what type of web design will each project require? e. what type of communication needs to exist between groups during the project? vi. concluding remarks vii. related links and resources cited viii. white paper appendixes a. working list of advisory committee functions and project workgroup functions b. contentdm software: roles and expertise c. project team workflow d. contentdm elements appendix c. first workshop questions general questions . how do you define a digital library? do the terms “repository,” “digital project,” “exhibit,” or “online collection” connote different things? if so, what are the differences, similarities, and boundaries for each? . what factors have contributed to a successful digitization program at your institution? did anything go drastically wrong? were there any surprises? what should new digitization programs be cautious and aware of? . what is the role, specifically, of the academic library in creating digital collections? how is digitization tied to the mission of your institution? . why digitize and for whom? do digital libraries need their own mission statement or philosophy because they differ from physical collections? should there be boundaries to what is digitized? . what standards are most widely in use at this time? what does the future hold? are there new standards you are interested in? technical questions, metadata questions . what are some of the recommended components of digital library infrastructure that should be in place to support a digitization program (equipment, staff, planning, technical expertise, content expertise, etc?) . what are the relationships between library digitization initiatives, the library website, the campus website or por- tal, and the web? in what ways do these information sources overlap, interoperate, or require boundaries? . how do you decide on what technology to use? what is the decision-making process when implementing a new technology? . standards are used in various ways during digitization. what is the importance of using standards, and are there areas where standards should be relaxed, or not used at all? how do digitization programs deal with evolving standards? . preservation isn’t talked about as much as it used to be. what’s your solution or strategy to the problem of preserv- ing digital materials? . will embedded metadata ever be the norm for digital objects, or will we continue to rely on collection management like contentdm to link digital objects to their associated metadata? information technology and libraries | september appendix d. second workshop outline . introduction—purpose/focus of the meeting a. to talk about next steps in the digitization program b. quick review of the current status and where the program has been c. serve to further educate participants on the steps involved in taking a project idea to reality d. goals for participants: understand types of projects and project prioritization; engage in activities on ideas and prioritization; talk about process and discuss committee; open forum . staff digitization survey discussion a. “defining digital libraries” b. “boundaries to the digitization program” c. “users and audience” d. “digital project design” e. “potential projects and ideas” . first group exercise: digital project idea ranking and defense of ranking . second group exercise: digital project idea brainstorming and defense of ideas brainstormed . concept/proposal for a digitization advisory committee . conclusion and next steps collections and design questions . how do you decide what should be included in a digital library? does the digital library need a collection develop- ment policy and if so, what type? how are projects prioritized at your institution? . how do you decide who your user is? are digital libraries targeting mobile users or other users with unique needs? what value-added material compliments and enhances digital collections (i.e., item-level metadata records, guided searches, narrative or scholarly content, teaching material, etc.)? . how should digital libraries be assessed and evaluated? how do you gauge the success of a digital collection, exhibit, or library? what has been proven and disproved in the short time that libraries have been doing digital projects? . what role do digital libraries play in marketing the library? how do you market your digital collections? are there any design criteria that should be considered for the web presence of digital libraries (should the digital library look like the library website, the campus website, or have a unique look and feel)? . do you have any experience partnering with teaching faculty to create digital collections? how are collabora- tions initiated? are such collaborations a priority? what other types of collaborations are you involved in now? how do you achieve consensus with a diverse group of collaborators? to what degree is centralization important or unnecessary? editorial | truitt marc truitteditorial a s i write this, hurricane ike is within twelve hours of making landfall in texas; currently, it appears that the storm will strike directly at the houston– galveston area. houstonians with long memories will be comparing ike to hurricane alicia, which devastated the region in , killing twenty-one and doing $ . billion in damage. younger residents and/or more recent immi- grants to the area will recall tropical storm allison, which though not of hurricane force, lashed the city and much of east texas for two weeks in june , leaving in its wake twenty-three dead, $ . billion in losses, and tens of thou- sands of homes damaged or destroyed. and of course, more recently, and much better known to all of us, regard- less of where we live, katrina, the “mother of all storms,” killed over eighteen hundred, caused over $ billion in damage, left huge swaths of new orleans uninhabitable, and created a population exodus with whose effects we are living even to this day. common to each of these disasters—and so many others like them—is the fact that they have often wrought terrible damage on libraries in their areas. most of us have probably seen the pictures of the water- and mildew- damaged collections at tulane, xavier, the university of new orleans, and the new orleans public library sys- tem. and the damage from these events is long-term or even permanent. i formerly worked at the university of houston (uh), and when i left there in that institu- tion was still dealing with the consequences of allison’s destruction of uh’s subterranean law library. and now i have to wonder whether uh librarians, faculty, and stu- dents might not be facing a similar or even worse catas- trophe all over again with ike. ital editorial board member donna hirst has done the profession a great service with her column, “the iowa city flood of : a librarian and it professional’s perspective,” which appears in this issue. her account of how library it folks there dealt with relocations of serv- ers, library staff, and indeed library it staff members themselves should be made required reading for all of us in the field, as well as for senior library administrators. the problem, i think we all secretly know, is that emer- gency preparedness—also known by its current moniker “business continuity planning” (bc)—and disaster recov- ery (dr) are not “sexy” subjects. devoting a portion of our always too modest resources of money, equipment, staffing, and time to what is, at best, a sort of insurance against what might happen someday seems inexcusably profligate today. such planning and preparation doesn’t roll out any shiny new services and will win few plaudits from staff or patrons, to say nothing of new resources from those who control our institutional purse strings. buying higher bandwidth equipment for a switching closet is likely to be a far easier sell. that is, until that unthinkable something happens, and your organization is facing (or suffers) a catastrophic loss of it services. note that i didn’t say “equipment” or “infrastructure.” the really important loss will be one of services. “stuff”—in the form of servers, workstations, net- works, etc.—all costs money, but ultimately is replaceable. what are not replaceable—at least not immediately—are library services to staff and patrons: access to comput- ing (networking, e-mail, productivity applications, etc.), internet resources, and perhaps most importantly nowa- days, the licensed electronic content on which we and our patrons have so come to rely. while the news coverage will emphasize (not without justice, i think) the lost or rescued books in a catastrophic loss situation, what staff and patrons are likely to demand first and loudest will be continuation or restoration of technology-based library services such as e-mail, web presence, web access, and licensed content. lest there be doubt, does anyone recall what drove evacuees into public libraries in the wake of katrina? it was, as much as anything, the desire to locate loved ones and especially the need to seek informa- tion and forms for government assistance—all of which required access to networked computing resources. if we have one at all—i suspect that many of us have a dr plan that is sadly dated and that has never been tested. look at it this way: would you roll out a critical and highly visible new web service without careful prep- aration and testing? yet many of us somehow think that bc or dr is somehow different, with no periodic review or testing required. since we feel we have no resources to devote to bc or dr planning and testing, we excuse our failure to do so by telling ourselves and our administra- tions that “we can’t really plan for a disaster, since the precise circumstances for which we’re planning won’t be the ones that actually occur.” and so we find ourselves later facing a crisis without any preparation. here at the university of alberta libraries, we’ve been giving the questions of business continuity and disaster recovery a good deal of thought lately. our preexisting dr plan was typical of the sort i’ve described above: out- of-date, vanishingly skeletal in its details, without explicit reference or relevance to maintenance and restoration of mission critical services, and of course, untested. impetus for our review has come from several sources. perhaps the most interesting of these has been a univer- sity-sponsored bc planning process that embraces a two- pronged approach: marc truitt (marc.truitt@ualberta.ca) is associate director, bibliographic and information technology services, university of alberta libraries, edmonton, alberta, canada, and editor of ital. information technology and libraries | december n identify and prioritize your organization’s services. working with other constituencies within the library, we have identified and prioritized approxi- mately ten broad services to be maintained or restored in the event of an interruption of our normal business activities. for example, our top priority is the continuation or restoration of access to licensed electronic content (e.g., e-journals, e-books, databases, etc.). our it disaster planning will be informed by and respond to this goal. n identify “upstream” and “downstream” dependencies. we are dependent on others for services so that we can provide our own; thus we cannot offer access to the internet for our users unless campus it provides us with a gateway to off-campus networks. we need to make certain as we plan that campus it is aware of and can provide this service in the scenarios for which we’re planning. by the same token, others are dependent on us for the provision of services critical to their planning: our consortial partners, for example, rely on us for ils, document delivery, and other technology-based services that we need to plan to continue in the event of a disaster. these two facets—services and dependencies—can be expressed as a matrix that is helpful in planning for bc and dr goals that are both responsive to the needs of the organization and achievable in terms of upstream and downstream dependencies. it has been an enlighten- ing exercise. one consequence has been our decision to include, as part of next fiscal year’s budget request, fund- ing to help create a dr site at our library’s remote storage facility, to enable us quickly to restore access to our most critical technology services. in the past, we might have used this annual request as an opportunity to highlight our need for funding to support rolling out some glamor- ous new service initiative. with this request, though, we are explicitly recognizing that we as an organization need to commit to measures that ensure the continuance in a variety of situations of our existing core services. that’s a major change in mindset for us, as i suspect it would be for many library it organizations. a final interesting aspect of our planning process is that one of the major drivers for the university is a con- cern about business continuity in the event of a people- based disaster. as avian influenza (aka, “bird flu”) has spread beyond the confines of its southeast asian point of origin, worry about how we continue to operate in the midst of a pandemic has been added to the more predict- able suite of fires, floods, tornadoes, and earthquakes (okay, not likely in alberta). indeed, pandemic planning is in many ways far more difficult than that for more “normal” disasters. while in many smaller libraries the “it shop” may be comprised of one person in many hats, in larger organizations such as ours (approximately full-time equivalent employees in library it), there tends to be a great deal of specialization. can the webmaster, in the midst of a crisis, support staff workstations? can the help desk technician deduce why our vendor for web of science has suddenly and inexplicably disabled our access? our bc process rules tell us that we should be planning for “three-deep” expertise in all critical areas, since the assumption is that a pandemic might mean that a third or more of our staff would be ill (or worse) at any given time. how many of us offer critical technology ser- vices that suffer from that it manager’s ultimate staffing nightmare, the single point of failure? we have no profound answers to these questions, and our planning process is by no means the one that will work for all organizations. but the evidence of katrina, ike, and iowa city is plain: we need to be as prepared as possible for these events. the time to “get religion” about business continuity and disaster recovery is before the unthinkable occurs, not after. are there any of you out there with experiences—either in preparation and plan- ning or in recovery operations—that you would consider sharing with ital readers? we all would benefit from your thoughts and experiences. i know i would! post-ike postscript. ike roared ashore four days ago and it is clear from media coverage since that galveston suffered a catastrophe and houston was badly damaged. reports from area libraries are sketchy and only today beginning to filter out. meanwhile, at the university of houston, the building housing the architecture library lost its roof, and the salvageable portions of its collection are to be relocated to the main m.d. anderson library. references . “hurricane alicia,” wikipedia, http://en.wikipedia.org/ wiki/hurricane_alicia (accessed sept. , ). . “tropical storm allison,” wikipedia, http://en.wikipedia .org/wiki/tropical_storm_allison (accessed sept. , ). . “hurricane katrina,” wikipedia, http://en.wikipedia .org/wiki/hurricane_katrina (accessed sept. , ). enhancing opac records for discover | griffis and ford patrick griffis and cyrus ford enhancing opac records for discovery this article proposes adding keywords and descriptors to the catalog records of electronic databases and media items to enhance their discovery. the authors contend that sub- ject liaisons can add value to opac records and enhance discovery of electronic databases and media items by providing searchable keywords and resource descriptions. the authors provide an examination of opac records at their own library, which illustrates the disparity of use- ful keywords and descriptions within the notes field for media item records versus electronic database records. the authors outline methods for identifying useful keywords for indexing opac records of electronic databases. also included is an analysis of the advantages of using encore’s community tag and community review features to allow subject liaisons to work directly in the catalog instead of collaborating with cataloging staff. a t the university of nevada las vegas (unlv) libraries’ discovery mini-conference, there was a wide range of initiatives and ideas presented. some were large-scale initiatives that focused on design- ing search platforms and systems as well as information architecture schemas that would enhance library resource discovery. but there was not much focus on enhancing the representation of library resources within the construct of bibliographic records in the opac. since searching plat- forms can only be as useful as the information available for searching, and since opac records are the method for representing the majority of library resources, we thought it important that the prominence of opac records and how they represent library resources be considered in the mini-conference. to that end, our presentation focused on enhancing the opac records for nonbook items to support their discoverability as opposed to focusing on search systems and information architecture schemas. our proposition was that subject liaisons’ expertise could be used to enhance opac records by including their own keyword search terms and descriptive summaries in opac records for electronic databases as well as records of media items. this proposition acts as a moderate approach to ini- tiatives that call for opac records to be opened for user- generated content in that this approach provides subject liaison mediation and expertise to modify records. as such, this approach may serve as an effective stopgap in cases where there is resistance toward permitting social tagging and user descriptions within opac records. such an initiative also is scalable, allowing liaisons to provide as few or as many terms as they want. such an initiative would require collaboration between cataloging staff and subject liaisons. n disparity between media and database records at unlv libraries, terms included in the notes fields of bibliographic records are indexed for keyword searching. in the case of media items, there is extensive use of notes to include descriptive terms that enhance discoverability for users. for example, notes for films indicate any awards the film has won as well as festivals in which it has been featured (see figure ). as a result, users can discover films through keyword searches of film awards or film festivals. a film student who is searching “cannes film festival” via a keyword search will generate results that include films owned by unlv libraries that have been featured at that festival. these keyword-searchable notes add value and discoverability for this type of material, and subject liaisons can be a source for such information. while it appears that notes in media records are heavily populated with a variety of user-centric informa- tion, there is relatively little use of descriptive notes for figure . the notes field in an opac record of a film item patrick griffis (patrick.griffis@unlv.edu) is business librarian and cyrus ford (cyrus.ford@unlv.edu) is special formats catalog librarian, university of nevada las vegas libraries. information technology and libraries | december electronic databases (see figure ). for databases, notes traditionally include information about access restrictions and mode of access while overlooking information rep- resenting the content of the resource. these fields could be utilized for specific terms relating to database con- tent not adequately covered by the library of congress subject headings (lcsh). subject liaisons have special- ized knowledge of which databases work best for unique content areas, class assignments, and information needs. this user-centric knowledge can be used to enhance database discovery if liaisons were to provide catalogers with information and descriptors to add to the record. as an example, at unlv libraries there is one par- ticular database that provides a strengths, weaknesses, opportunities, and threats (swot) analysis for compa- nies, but that natural language term isn’t found anywhere in the general database summary listing or subject head- ings. if it were added to a note field as part of a descrip- tion or as a labeled descriptor, then students could easily find this database to complete their assignments. this proposal is scalable, allowing liaisons to provide as few or as many key terms as they want, depending on their preference or on the vagaries of a particular data- base. subject liaisons could opt to add a few major terms from their own knowledge and expertise that they feel will add value for patrons searching the opac. subject liaisons also could mine the index and thesaurus terms of individual databases to identify prominent content areas for individual databases to find useful keywords. n mining electronic database index descriptors electronic databases typically have subject matter tax- onomies developed by experts who assign descriptors to journal articles. subject liaisons could mine these taxonomies to identify predominant descriptors for indi- vidual databases to add to the database catalog records. predominance of a subject descriptor could be deter- mined by examining the relative number of articles that are assigned to that descriptor. such a strategy of indexing key predominant subject descriptors identified from database subject matter taxonomies could serve to uncover unique content areas not served with lcsh. a different application of this strategy could be employed for identifying predominant and emerging research areas for particular groups. subject liaisons could conduct a citation analysis of articles authored by members of a particular research group to record and codify the subject descriptors of each article. once codi- fied, an analysis could determine the most predominant subject descriptors for articles authored by that particu- lar group. this could serve as a baseline for identifying emerging research areas and their terms. both types of analysis have potential to provide useful keyword terms for database records. n using encore’s community features in , unlv libraries purchased and implemented the innovative interfaces’ encore discovery platform, which provides a google-like interface for searching the public catalog and the ability to narrow results using fac- ets such as location, year, language, and format. encore also includes many display features that showcase the information provided in the bibliographic records. two of encore’s web . features provide users with the abil- ity to contribute data to records via community tags and community reviews. unlv requires users to enter a valid library barcode number and pin. subject liaisons could use the community reviews feature to add descriptive summaries of items to encore records independently, without the need for cataloging staff to edit a marc record. however, the content of community reviews are not indexed for searching and thus only add value at the point when a user is determining whether the resources they have retrieved are valuable for them. on the other hand, if a community tag is added to an item, that tag is included in the community tags section figure . the notes field in an opac record of an electronic database enhancing opac records for discover | griffis and ford of the encore result display and becomes an indexed keyword for searches in encore (see figure ). if that tag term is searched in encore’s keyword search, the biblio- graphic record attached to that tag term will be included in the results list under the community tags facet. since these community tags are searchable, subject liaisons can add keywords to encore records without collaboration with cataloging staff. however, this provides limited suc- cess because the keyword is included and indexed only in encore records—not in the opac records. also, the community tags facet must be selected from the results display for the encore record tags to be searchable. n the case for collaboration as described above, keywords and descriptions added by subject liaisons into encore records have inherent dis- covery limitations when compared to a cataloger adding the same information directly to the marc bibliographic record. the advantages of collaboration between subject liaisons and catalogers is clear, and subject librarians at unlv libraries have experienced similar collaboration in efforts in the past. in , subject librarians at unlv libraries were offered the opportunity to create their own descriptions of electronic resources through an initiative to update the summary descriptions for the electronic databases portion of the libraries’ website. at that time, all existing electronic database summaries were those used by the database publishers. the project provided subject liaisons the option to create custom summary descriptions to rep- resent electronic databases in their own terms. each sub- ject liaison had a document file for their descriptions, and the website editors used them to update the electronic databases list on the libraries’ website. this particular initiative serves as one example of the willingness of subject liaisons to share their subject exper- tise to enhance the representation of library resources through collaboration with technical services staff. as such, collaboration between subject liaisons and catalog- ers to allow liaisons to add terms to opac records of electronic databases and media items could prove to be both effective and feasible as an initiative toward enhanc- ing the discovery of library resources. figure . encore community tag lita cover , cover , cover index to advertisers editorial board thoughts | eden editorial board thoughts bradford lee eden musings on the demise of paper w e have been hearing the dire predictions about the end of paper and the book since microfiche was hailed as the savior of libraries decades ago. now it seems that technology may be finally catching up with the hype. with the amazon kindle and the sony reader beginning to sell in the marketplace despite the cost (about $ for the kindle), it appears that a whole new group of electronic alternatives to the print book will soon be available for users next year. amazon reports that e-book sales quadrupled in from the previous year. this has many technology firms salivating and hop- ing that the consumer market is ready to move to digital reading as quickly and profitably as the move to digital music. some of these new devices and technologies are featured in the march , , fortune article by michael v. copeland titled “the end of paper?” part of the problem with current readers is their chal- lenges for advertising. because the screen is so small, there isn’t any room to insert ads (i.e., revenue) around the margins of the text. but new readers such as plastic logic, polymer vision, and firstpaper will have larger screens, stronger image resolution, and automatic wire- less updates, with color screens and video capabilities just over the horizon. still, working out a business model for newspapers and magazines is the real challenge. and how much will readers pay for content? with everything “free” over the internet, consumers have become accus- tomed to information readily available for no immediate cost. so how much to charge and how to make money selling content? the plastic logic reader weighs less than a pound, is one-eighth of an inch thick, and resembles an ½ x inch sheet of paper or a clipboard. it will appear in the mar- ketplace next year, using plastic transistors powered by a lithium battery. while not flexible, it is a very durable and break-resistant device. other e-readers will use flexible display technology that allows one to fold up the screen and place the device into a pocket. much of this technol- ogy is fueled by e-ink, a start-up company that is behind the success of the kindle and the reader. they are explor- ing the use of color and video, but both have problems in terms of reading experience and battery wear. in the long run, however, these issues will be resolved. expense is the main concern: just how much are users willing to pay to read something in digital rather than analog? amazon has been hugely successful with the kindle, selling more than , for just under $ in . and with the drop in subscriptions for analog magazines and news- papers, advertisers are becoming nervous about their futures. or will the “pay by the article” model, like that used for digital music sales, become the norm? so what should or do these developments mean for libraries? it means that we should probably be exploring the purchase of some of these products when they appear and offering them (with some content) for checkout to our patrons. many of us did something similar when it became apparent that laptops were wanted and needed by students for their use. many of us still offer this ser- vice today, even though many campuses now require students to purchase them anyway. offering cutting-edge technology with content related to the transmission and packaging of information is one way for our clientele to see libraries as more than just print materials and a social space. and libraries shouldn’t pay full price (or any price) for these new toys; companies that develop these products are dying to find free research and devel- opment focus groups that will assist them in versioning and upgrading their products for the marketplace. what better avenue than college students? related to this is the recent announcement by the university of michigan that their university press will now be a digital operation to be run as part of the library. decreased university and library budgets have meant that university presses have not been able to sell enough of their monographs to maintain viable business models. the move of a university press to a successful scholarly communication and open-source publishing entity like the university of michigan libraries means that the press will be able to survive, and it also indicates that the newer model of academic libraries as university publishers will have a prototypical example to point out to their univer- sity’s administration. in the long run, these types of part- nerships are essential if academic libraries are to survive their own budget cuts in the future. references . michael v. copeland, “the end of paper?” cnnmoney .com, mar. , , http://money.cnn.com/ / / / technology/copeland_epaper.fortune/ (accessed june , ). . andrew albanese, “university of michigan press merged with library, with new emphasis on digital mono- graphs,” libraryjournal.com, mar. , , http://www .libraryjournal.com/article/ca .html (accessed june , ). bradford lee eden (eden@library.ucsb.edu) is associate university librarian for technical services and scholarly communication, university of california, santa barbara. information technology and libraries | june mark beatty (mbeatty@wils.wisc.edu) is lita president / and trainer, wisconsin library services, madison. mark beattypresident’s message i’ve recently read three quite different articles that surprisingly all had something similar to say with a different twist on the theme uppermost in my brain the last year or two. here’s the briefest of quotes from the three. i would suggest your full reading of all three if you haven’t already. n lankes, silverstein and nicholson, “participatory networks: the library as conversation,” in the december information technology and libraries: “with their principles, dedication to service, and unique knowledge of infrastructure, libraries are poised not simply to respond to new technologies, but to drive them. by tying technological implementation, development and improvement to the mission of facilitating conver- sations across fields, libraries can gain invaluable visibility and resources.” n bill crowley, “lifecycle librarianship,” in the april , library journal: “public, academic, and school librarians should adopt the service philosophy of lifecycle librarianship and jointly plan at town, city, or county levels to identify and meet human learning needs from “lapsit to nursing home.”” n joe kissell, “instant messaging for introverts,” in the april , tidbits (http://db.tidbits.com/ article/ ): “several people i discussed this issue with (using im and twitter) expressed dismay at having had relationships deteriorate due to an unwillingness on another person’s part to adapt to changing technology. for example, people who don’t use e-mail don’t get evites, and so they end up being excluded from parties.” what all three express to me is a concern that librar- ies, and just plain humans, need to be part of the conver- sation, part of the social structure, and full participants in life. we are now, through surveys and meetings and focus groups, starting to know that new librarians and new lita members are most interested in networking with their colleagues using multiple methods to fulfill the whole range of their professional and social needs. lankes wants to make sure we participate with all our constituencies, crowley wants us to spend a lifetime with those constituencies, and kissell wants to make sure we get invited to the party. that sounds a bit face- tious but i believe the point is that our association, our libraries, our social structures are now required to be active participants, physically and virtually, in the life of their communities. we have to recognize our communi- ties and then act to participate and provide space and support to those communities. this takes work and the will to always be part of our communities. all of which leads to my president’s program, featur- ing keynote speaker joe janes and the blogging folks at “it’s all good” at the ala annual conference in anaheim, california. it will be part of sunday afternoon with lita, taking place on sunday, june , . the program line up will include: n top technology trends : – : p.m. n lita awards and scholarships reception – p.m. n lita president’s program – p.m. n “isn’t it great to be in the library . . . wherever that is?” it’s often said that today we have to run three libraries at once: the library of yesterday, today, and tomorrow. we run both the physical, visible library, and the one that exists beyond the walls. this raises many questions of what a library is and encompasses, what it isn’t, where the boundaries lie, the impact on what we do and how we do it, what our clients want, how we serve them, and what kinds of librarians serve them. this program will attempt to examine the full social and cultural constructs of libraries that move beyond basic web . and integrate patrons, librarians, and resources in what should be a ubiquitous manner. join joe janes, associate professor in the information school of the university of washington in seattle and columnist for american libraries, keynote speaker, along with members of the “it’s all good” blog- ging group (http://scanblog.blogspot.com) as the reactor panel for a lively exploration of possible futures. i hope you’ll be able to attend but be assured that members of the lita community will blog and report and even record the sessions in various ways that will be made freely available to our community. information technology and libraries | june information technology and libraries | september from our readers: virtues and values in digital library architecture mark cyzyk editor’s note: “from our readers” will be an occasional feature, highlighting ital readers’ letters and commen- taries on timely issues. at the fall coalition for networked information (cni) conference in washington, d.c., i pre-sented “a survey and evaluation of open-source electronic publishing systems.” toward the end of my pre- sentation was a slide enumerating some of the things i had personally learned as a web application architect during my review of the systems under consideration: n platform independence should not be neglected. n one inherits the flaws of external libraries and frameworks. choose with care. n installation procedures must be simple and flawless. n don’t wake the sysadmin with “slap a gui on that xml!”—and push application administration out, as much as possible, to select users. n documentation must be concise, complete, and comprehensive. “i can’t guess what you’re thinking.” initially, these were just notes i thought might be useful to others, figuring it’s typically helpful to share experiences, especially at international conferences. but as i now look at those maxims, it occurs to me that when abstracted further they point in the direction of more general concepts and traits—concepts and traits that accurately describe us and the products of our labor if we are successful, and prescribe to us the concepts and traits we need to understand and adopt if we are not. in short, peering into each maxim, i can begin to make out some of the virtues and values that underlie, or should underlie, the design and architecture of our digital library systems. n freedom and equality platform independence should not be neglected. “even though this application is written in platform- independent php, the documentation says it must be run on either red hat or suse, or maybe it will run on solaris too, but we don’t have any of these here.” while i no doubt will be heartily flamed for suggest- ing that microsoft has done more to democratize comput- ing than any other single company, i nevertheless feel the need to point out that, for many of us, windows server operating systems and our responsibility for adminis- tering them way back when provided the impetus for adding our swipe-card barcodes to the acl of the data center—surely a badge of membership in the club of enterprise it if ever there was one. you may not like the way windows does things. you may not like the way microsoft plays with the other boys. but to act like they don’t exist is nothing more than foolish burying one’s head in the *nix sand. windows servers have proven themselves time and again as being affordable, easily managed, dependable, and, yes, secure workhorses. windows is the ford pickup truck of the server world, and while that pickup will some day inevitably suffer a blowout of its twenty-year-old head gasket (and will therefore be respectfully relegated to that place where all dearly departed trucks go), it’s been a long and good run. we should recognize and appreciate this. windows clearly has a place in the data center, sitting quietly humming alongside its unix and linux brothers. i imagine that it actually takes some effort to produce platform-dependent applications using platform-inde- pendent languages and frameworks. such effort should be put toward other things. keep it pure. and by that i mean, keep it platform independent. freedom to choose and presumed equality among the server-side oses should reign. n responsibility and good sense one inherits the flaws of external libraries and frame- works. choose with care. so you’ve installed the os, you’ve installed and configured the specified web server, you’ve installed and configured the application platform, you’ve downloaded and com- piled the source, yet there remains a long list of external libraries to install and configure. one by one you install them. suddenly, when you get to library number you hit a snag. it won’t install. it requires a previous version of library number , and multiple versions of library number can’t be installed at the same time on the same box. worse yet, as you take a break to read some more of the documentation, it sure looks like required library number is dependent on the current version of library number and won’t work with any previous version. and could it be that library number is dependent on library number , which is dependent on library number , which is dependent on—yikes—library number ? mark cyzyk (mcyzyk@jhu.edu) is the scholarly communication architect, library digital programs group, sheridan libraries, johns hopkins university in baltimore. from our readers: virtues and values in digital library architecture | cyzyk all things come full circle. but let’s suppose you’ve worked out all of these dependencies, you’ve figured out the single, secret order in which they must install, you’ve done it, and it looks like it’s working! yet, when you go to boot up the web service, sud- denly there are errors all over the place, a fearsome crash- ing and burning that makes you want to go home and take a nap. something in your configuration is wrong? something in the way your configuration is interacting with an external library is wrong? you search the logs. you gather the relevant messages. they don’t make a lot of sense. now what to do? you search the lists, you search the wikis to no avail, and finally, in desperation, you e-mail the developers. “but that’s a problem with library x, not with our application.” au contraire. i would like to strongly suggest a copernican revolu- tion in how we think about such situations. while it’s obvious that the developers of the libraries themselves are responsible for developing and maintaining them, i’d like to suggest that this does not relieve you, the developer of a system that relies on their software, from responsi- bility for its bugs and peculiar configuration problems. i’d like to suggest that, far from pushing responsibility in the case mentioned above out to the developers of the malfunctioning external library, that you, in choos- ing that library in the first place, have now inherited responsibility for it. even if you don’t believe in this notion of inheritance, if you would please at least act as if it were true, we’d all be in a better place. part of accepting this kind of respon- sibility is you then acting as a conduit through which we poor implementers learn the true nature of the problem and any solutions or temporary workarounds we may apply so that we can get your system up and running pronto. in the end, it’s all about your system. your system as a whole is only as strong as the weakest link in its chain of dependencies. n simplicity and perfection installation procedures must be simple and flawless. it goes without saying that if we can’t install your system we a fortiori can’t adopt it for use in our organization. i remember once having such a difficult time trying to get a system up and running that i almost gave up. i tried first to get it running against apache . , then against apache . . i had multiple interactions with the develop- ers. i banged my head against the wall of that system for days in frustration. the documentation was of little help. it seemed to be more part of an internal documentation project, a way for the developers to communicate among themselves, than to inform outsiders like me about their system. and related to this i remember driving to work during this time listening to a report on npr about the famous hopkins pediatric neurosurgeon, dr. ben carson. apparently, earlier in the week he had separated the brains of siamese twins and the twins were now doing fine, recuperating. the npr commentator marveled at the intricacy of the operation and at the fact that the whole thing took, i believe, five hours. “five hours? five hours?!” i exclaimed while bar- reling down the highway in my vintage ford ranger pickup (head gasket mostly sealed tight, no compression leakage). “i can’t get this system at work installed in five days!” our goal as system architects needs to be that we provide to our users simple and flawless installation pro- cedures so that our systems can, on average, be installed and configured in equal or less time than it takes to per- form major brain surgery. “all in an afternoon” should become our motto. i am happy to find that there are useful and easy to use package managers, e.g., yum and synaptic, for doing such things on various linux distributions. windows has long had solid and sophisticated installation utilities. tomcat supports drop-in-place war files. when possible and appropriate, we need to use them. n justice and e-z livin don’t wake the sysadmin with “slap a gui on that xml!”—and push application administration out, as much as possible, to select users. i remember reading plato’s republic as an undergraduate and the feeling of being let down when the climax of the whole thing was a definition in which “justice” simply is each man serving his proper place in society and not transgressing the boundaries of his role. “that’s it?” i thought. “so you have this rigidly hier- archical society and each person in it knows his role and knows in which slot his role fits—and keeping to this is ‘justice’?” this may not be such a great way to structure a soci- ety, but now that i think about it, it’s a great way to struc- ture a computer application. sit down and carefully look at the functions your program will provide. then create a small set of user roles to which these functions will be carefully mapped. in the end you will have a hierarchical structure of roles and functions that should look perfectly simple and rational when drawn on a piece of paper. and while the superuser role should have power over information technology and libraries | september all and access to all functions in the application, the list of functions that he alone has access to should be small, i.e., the actual work of the superuser should be minimized as much as possible by making sure that most functions are delegated to the members of other, appropriate, proper user roles. doing this happily results in what i call the state of e-z livin: the last thing you want is for users to con- stantly be calling you with data issues to fix. you there- fore will model management of the data—all of it—and the configuration of the application itself—most of it— directly into the architecture of the application, provide users the guis they need to configure and manage things themselves, and push as much functionality as you can out to them where it belongs. let them click their respec- tive ways to happiness and computing goodness. you build the tool, they use it, and you retire back to the land of e-z livin. users are assigned to their roles, and all roles are in their proper places. application architecture justice is achieved. n clarity and wholeness documentation must be concise, complete, and compre- hensive. “i can’t guess what you’re thinking.” as system developers we’ve probably all had the magical experience of a mind meld with a fellow developer when working intensively on a project. i have had this experi- ence with two other developers, separately, at different stages of my career. (one of them, in fact, used to point out to everyone that, “between the two of us, we make one good developer!”) this is a wonderful and magical and productive working relationship in which to be, and it needs to be recognized, supported, and exploited whenever it happens. you are lucky if you find yourself designing and developing a system and your counterpart is reading your mind and finishing your sentences. however, just as it’s best to leave that nice young couple cuddling in the corner booth alone, so too it really doesn’t make a lot of sense to expect the mind-melded developers to turn out anything that remotely resem- bles coherent and understandable documentation. those undergoing a mind meld by definition know perfectly well what they mean. to the rest of us it just feels like we missed a memo. if you have the luxury, make sure that the one writ- ing the documentation is not currently undergoing a mind meld with anyone else on the development team. scotty typically stayed behind while he beamed the oth- ers down. beam them down. be that scotty. you do the world a great service by staying behind on the ship and dutifully reporting, clearly and comprehensively, what’s happen- ing down on the red planet. to these five maxims, and their corresponding vir- tues, i would add one more set, one upon which the others rely: n empathy and graciousness you are not your audience. at least in applied computing fields like ours, we need to break with the long-held “guru in the basement” mentality. the actions of various managerial strata have now ostensibly acknowledged for us that technical exper- tise, especially in applied fields, is a commodity, i.e., it can be bought. a dearth of such expertise is remedied by sim- ply applying money to the situation—admittedly difficult to do at the majority of institutions of higher education, but a common occurrence at the wealthiest. nevertheless, the dogmatic hold of the guru has been broken and the magical aura that once draped her is not now so resplen- dent—her relative rarity, and the clubby superiority that depended upon it, has been diluted significantly by the sheer number of counterparts who can and will gleefully fill her function. we respect, value, and admire her; it’s just that her stranglehold on things has (rightfully) been broken. and while nobody is truly indispensable, what is more difficult and rare to find is someone who has the guru’s same level of technical chops coupled with a genuine empathic ability to relate to those who are the intended users of her systems and services. unless your systems and services are geared primarily toward other developers, programmers, and architects— and presumably they are not, nor, in the library world, should they be—your users will typically be significantly unlike you. let me repeat that: your users are not like you. rephrased: you are not your audience. when looking back over the other maxims, values, and virtues mentioned in this essay then, the moral- psychological glue that binds them all is composed of empathy for our users—faculty, students, librarians, non-technical staff—and the graciousness to design and carry out a project plan in a spirit of openness, caring, flexibility, humility, respect, and collaboration. when empathy for the users of our systems is absent—and there are cases where you can actually see this in the design and documentation of the system itself—our systems will ultimately not be used. when the spirit of graciousness is broken, men become robots, mere rule followers, and users will boycott using their systems and will look else- from our readers: virtues and values in digital library architecture | cyzyk where, naturally preferring to avoid playing the simon- says games so often demanded by tech folk in their workaday worlds; there is a reason the comic strip dilbert is so funny and rings so true. when confronted with a lack of empathy and graciousness on our part, the users who can boycott using our systems and services will boy- cott using our systems and services. and we’ll be left out in the rain, feeling like, as bonnie raitt once sadly sang, “i can’t make you love me if you don’t / i can’t make your heart feel something it won’t.” empathy and gracious- ness, while not guaranteeing enthusiastic adoption of our systems and services, are a necessary precondition for users even countenancing participation. there are undoubtedly other virtues and values that can usefully be expounded in the context of digital library architecture—consistency, coherence, and elegance imme- diately come to mind—and i could go on and on analyz- ing the various maxims surrounding these that bubble up through the stack of consciousness during the course of the day. yet doing so would conflict with another virtue i think is key to the success and enjoyment of opinion- piece essays like this and maybe even of other sorts of publications and presentations: brevity. note . a colleague of mine has since informed me that car- son’s operation took twenty-five hours, not five. nevertheless, my admonition here still holds. when installation and con- figuration of our systems are taking longer, significantly longer, than it takes to perform major brain surgery, surely there is something amiss? author id box for column layout editorial board thoughts | hirst donna hirsteditorial board thoughts the iowa city flood of : a librarian and it professional’s perspective d o you like to chase fire trucks? do you enjoy watch- ing a raft of adventurers go over the waterfall, careening from rock to rock? well, this is a story of the iowa city flood of , a flood projected to happen once every five hundred years, from the perspective of a librarian and it professional. n the approach of the flood the winter of was hard, and we got mounds of snow. the spring was wet that year in iowa city. it rained almost every day. minnesota’s snow melt-off hadn’t been released from the reservoir due to the heavy rains. everyone watched the river rise, day by day. the parks were underwater; the river was creeping up toward buildings, including the university of iowa. in early june, with about a day and a half notice, library staff at the university’s main library, art library, and music library were told to evacuate. one of the first acts of evacuation was the relocation of all of the library servers to the engineering building up the hill—high and dry—literally rolling them across the street and up the sidewalk. although all servers were relocated to engi- neering, engineering didn’t have enough power in their server room to handle the extra capacity to run all of our machines. the five primo servers that run our discovery searching service had to stay disconnected. with the servers safe and sound, we moved our atten- tion to staff workstations. the personal workstations of the administrative staff and the finance department were moved to the business library. the libraries’ laptops were collected and moved into the branch libraries, which would be receiving displaced staff. many staff would be expected to work from public clusters in the various library branches, locked down to specific functions. as library staff were collecting their critical posses- sions, the town was madly sandbagging. more than a million sandbags were piled around university buildings, private businesses, and residences. in retrospect, some of the sandbags may have made a difference, but since the flood was so much greater than anticipated, the water largely went over and around, leaving a lot of soggy sandbags. on june , the day before the main library was to be closed, the decision was made to move books up from the basement. there were well over , volumes in the basement, and a group of approximately five hundred volunteers moved , volumes and , manuscript boxes from the lower shelves. volunteers passed books hand to hand into the third, fourth, and fifth floors of the building. a number of the volunteers came from sandbagging teams. individuals who had never been in a boxes of manuscripts being stacked on the fifth floor photo by carol jonck moving boxes out of the basement photo courtesy of the university of iowa news service donna hirst (donna-hirst@uiowa.edu) is project coordinator, library information technology, university of iowa libraries, iowa city. information technology and libraries | december library, didn’t know what a circulation desk was, or what a library of congress call number was were working hard side by side with physicians, ministers, scientists, students, and retirees. the end result was not orderly, but the collection was saved from the encroaching river. the libraries at the university of iowa are indebted to these volunteers who helped protect the collection from the expected water. n the river peaks approximately twenty university buildings were closed because of the flood, including the main library, the art building, and the music building. the university’s power plant was closed. the entire arts campus was deeply under water. most of the main roads connecting the east side of iowa city to the west side were closed, and most of the highways into iowa city were closed. interstate was closed in multiple places, and no traffic was allowed from the east side of the state to the west side. many bridges in and around iowa city were closed; some had actually crumbled and floated down stream. so the president of the university, sally mason, closed the university for the first time in its history. most staff would not be able to get to work anyway. many individu- als were struggling with residences and businesses that were under water. the university was to be closed for the week of june , with the university’s hospitals con- tinuing to operate under strained conditions; continued delivery of patient services was a priority. most library staff stayed home and followed the news stories, shocked at the daily news of destruction and loss. select library it staff began working in the background to set up new work environments for library staff returning to foreign workstations or relocated work environments. at the flood’s peak, the main library took several inches of water in the basement. there was slight rust- ing in the compact shelving, but the collection was com- pletely saved. a portion of the basement was lower, and the computer equipment controlling the libraries’ public computer cluster was completely ruined. this computer cluster housing more than two hundred workstations library staff and volunteers sandbagging photo by carol jonck moving books out of the basement photo courtesy of the university of iowa news service the beginning of a book chain to the fourth floor photo courtesy of the university of iowa news service editorial board thoughts | hirst which had been moved on the last day before the evacu- ation. much of this administrative work could proceed, and during the first week at the business library our finance department successfully completed our end-of- year rollover process on all our materials funds. staff from the music library, art library, preservation, and spe- cial collections were assigned to the business library. the engineering library adopted the main library circulation and reserve departments. the media services staff was relocated to the physics library. the media staff had cleverly pulled most of the staff development videos and made them available to staff from the physics library, thus allowing the many displaced library staff to make progress on staff develop- ment requirements. was completely out of commission. the basements and first floors of the art and music buildings were completely ruined, but the libraries for these disciplines were on higher floors. the collections were spared, but there was absolutely no access to the building. n the libraries take baby steps to resume service after a week of being completely shut down, the university opened to a first day of summer school, but things were not the same. for the nineteen university buildings that had been flooded, hordes of contractors, subcontractors, and labor- ers began the arduous task of reclamation. university staff could work at home when that was possible, and most of the library’s dislocated reference staff did that, devel- oping courses for the fall, progressing on selection work, and so on. staff could take vacation, but few chose this option. approximately staff from the main library and the art and music libraries were reassigned to four branch libraries that were not affected by the flood. all of central technical services (cts) and interlibrary loan staff were assigned to the hardin health science library. central shipping and facilities was also at harden library, thus the convoluted distribution of mail started from here. most of the pub- lic machines were taken by cts staff, but their routine work proceeded very slowly. cts did not have access to oclc until the end of their flood relocation, which seri- ously impacted their workflow. an early problem that had to be solved was providing telephones and printing to relocated staff. virtually none of the relocated staff had dedicated telephones, even the administration. in any given location the small number of regular branch staff graciously shared their phones with their visitors. sharing equipment tended to be true for printers as well. for a few critical phone numbers in the main library, the phone number was transferred to a designated phone in the branch. thus often, when regu- lar staff or student workers answered a phone, they had no idea what number the originating caller was trying to call. staff were encouraged to transfer their office phone number to their cell phone. at the business library, the library administrative staff and the finance staff had their personal workstations, library staff sandbagging photo by donald baxter information technology and libraries | december was closed for about four weeks. the art and music libraries may be closed for a year. when library staff returned to the main library, there were books and man- uscript boxes piled on the floor and on top of all the study tables. some of the main corridors, approximately twenty- one feet wide, were so filled with library materials that you almost had to walk sideways and suck in your tummy to walk down the hall. bathrooms were blocked and access to elevators was lim- ited. every library study table on the third through fifth floors were piled three feet high or more with books. for many weeks, library staff and volunteers care- fully sorted through the materials and reshelved them as required. many mate- rials needed conservation treatment, not because of the flood, but because of age and handling. many adjustments needed to be made to resume full service. due dates for all circula- tion categories had to be retrospectively altered to allow for the libraries being closed and for the extraordinary situations in which our library users found themselves during the flood. library materials were returned wet and moldy, and some items were lost. during the flood, in some cases, buildings actually floated down river. the libraries’ preservation department did extensive commu- nity education regarding treatment of materials damaged in the flood. the university was very interested in documenting the affect of the flood, and thus the libraries cooperated in trying to gather statistics on the number of hours of library staff and volunteers used during the flood. record keeping was complex, since one person could be a staff person working on flood efforts but also a volunteer working evenings and weekends. n our neighbors the effect of the iowa city flood of has been exten- sive, but was nothing compared to the flood in cedar rapids, our neighbor to the north. the cedar rapids public library lost their entire collection of , vol- umes, except for the children’s collection and , vol- umes that were checked out to library users that week. it staff were housed throughout the newly distrib- uted libraries complex. one it staff member was at the engineering library, one was at the health science library, and two were at the business library. several it staff were relocated to the campus computer center. n the libraries proceed apace despite hurdles as the water receded and workers cleaned and pro- ceeded with air handling and mold abatement, a very limited number of library staff were allowed back into the main library, typically with escorts, for very limited periods of time. during this time it staff was able to go into the main library and retrieve barcode scanners to allow cts staff to progress with book processing. staff went back for unprocessed materials needing original cataloging since staff had the time to process materials but didn’t have the materials. it staff retrieved some of our zebra printers so that labels could be applied to unbound serials. as it staff were allowed limited access to the main library, they went around to the various staff workstations and powered them up so that relocated staff could utilize the remote desktop function. n moving back the art and music libraries were evacuated june . the main library was evacuated june . the main library passing the books up the stairs photo courtesy of the university of iowa news service introducing zoomify image | smith column title editor author id box for column layout playing tag in the dark: diagnosing slowness in library response time | brown-sica margaret brown-sicatutorial playing tag in the dark: diagnosing slowness in library response time in this article the author explores how the systems department at the auraria library (which serves more than thirty thousand primarily com- muting students at the university of colorado–denver, the metropolitan state college of denver, and the community college of denver) diag- nosed and analyzed slow response time when querying proprietary databases. issues examined include vendor issues, proxy issues, library network hardware, and bandwidth and network traffic. w hy is everything so slow?” this is the question that library systems depart- ments often have the most trouble answering. it is also easy to dismiss because it is often the fault of factors beyond the control of library staff. what usually prompts these ques- tions are the experiences of the refer- ence librarians. when these librarians are trying to help students at the reference desk, it is very frustrating when databases seem to respond to queries slowly, files take forever to load onto the computer screen, and all the while the line in front of the desk get continues to grow. or the library gets calls from students using databases and the catalog from their homes who complain that searching library resources takes too long, and that they are getting frustrated and using google instead. this question is so painful because libraries spend so much of their shrinking budgets on high quality information in the form of expensive proprietary databases, and it is all wasted if users have trouble using them. in this case the problem seemed to be how slow the process of searching for information and downloading documents from databases was. for lack of a better term, the auraria library called this the “response time” problem. this article will discuss the various ways the systems (technology) department of the auraria library, which serves the university of colorado–denver, metropolitan state college of denver, and the community college of denver, tried to identify problems and improve database response time. the systems department defined “response time” as the time it took for a person to send a query from a computer at home or in the library to a proprietary information database and receive a response back, or how long it took to load a selected full- text article from a database. when a customer sets out to use a database in the library, the query to the database could be slowed down by many dif- ferent factors. the first is the proxy, in our case innovative interfaces’ inc. web access management (iii wam), a product that authenticates the user via the iii api (application program interface) product. to do this the query travels over network hardware, switches, and wires to the iii server and back again. then the query goes to the database’s server, which may be almost anywhere in the world. hardware problems at the database vendor’s end can affect this transfer. in the case of auraria library this transfer can be influenced by traffic on the library’s network, the university’s network, and any other place in between. this could also be hampered by the amount of memory in the computer where the query originates, by the amount of tasks being performed by that computer, etc. the bandwidth of the network and its speed can also have an effect. basically, the bottlenecks needed to be found and fixed. bottlenecks are described by webopedia as “the delay in transmission of data through the circuits of a computer’s micro- processor or over a tcp/ip network. the delay typically occurs when a system’s bandwidth cannot support the amount of information being relayed at the speed it is being pro- cessed. there are, however, many factors that can create a bottleneck in a system.” literature review there is not a lot on database response slowness in library literature, prob- ably because the issue overlaps with computer science and really is not one problem but a possibility of one of several problems. the issue is figuring out where the problem lies. gerhan and mutula examined tech- nical reasons for network slowness, performing bandwidth testing at a library in botswana and one in the united states using the same com- puter, and giving several suggestions for testing, fixing technical problems, and issues to examine. gerhan and mutula concluded that bandwidth and insufficient network infrastruc- ture were the main culprits in their sit- uation. they studied both bandwidth and bandwidth “squeeze.” looking for the bandwidth “squeeze” means looking along the internet’s “journey of many stages through routers and exchange points, each successively farther removed from the user.” bandwidth bottlenecks could occur at any one or more of those stages in the query’s transmission. the following four sections parse that lengthy path- way and examine how each may con- tribute to delays. badue et al. in their article “basic issues on the processing of web queries,” described web margaret brown-sica (margaret.brown -sica@ucdenver.edu) is head of technology and distance education support, auraria library, serving the university of colorado–denver, metropolitan state college of denver, and the community college of denver. information technology and libraries | december information technology and libraries | december queries, load balancing, and how they function. bertot and mcclure’s “assessing sufficiency and quality of bandwidth for public libraries” is based on data collected as part of the public libraries and the internet study and provides a very straight- forward approach for checking spe- cific areas for problems. it outlines why basic data such as bandwidth readings may not give the complete picture. it also gives a nice outline of factors involved such as local settings and parameters, ultimate connectivity path, application resource needs, and protocol priority. azuma, okamoto, hasegawa, and masayuki’s “design, implementation and evaluation of resource management system for internet servers” was very helpful in understanding the role and function of proxy servers and problems they can present. vendor issues this is a very thorny topic because it is out of the library’s control, and also because the library has so many data- bases. the systems department asked the reference staff to send reports of problems listing the type of activity attempted, time and dates, the names of the database, the problem and any error messages encountered. a few that seemed to be the slowest were selected for special examination. one vendor worked extensively with the library and in the end it was believed that there were problems at their end in load balancing, which eventually seemed to be fixed. that company was in the middle of a merger and that may have also been an issue. we also noted that a database that uses very large image files, artstor, was hard to use because it was so slow. this company sent the library an appli- cation that simulated the databases’ use and was supposed to test to see if bandwidth at auraria library was sufficient for that database. according to the test, it was. databases that con- sistently were perceived as the slowest were those that had the largest docu- ments and pictures, such as those that used primarily pdfs and visual material. this, with the results of the testing, pointed to a problem indepen- dent of vendor issues. bandwidth and network traffic the systems department decided to do bandwidth testing on the library’s public and staff computers after read- ing gerhan and mutula’s article about the university of botswana. the gen- eral perception is that bandwidth is often the primary problem in net- work slowness, as well as the prob- lems with databases that use larger files. several of the computers were tested in several successive days dur- ing what is usually the busiest time for the network, between noon and p.m. the results were good, averag- ing about kilobytes per second (kbps). for this test we used the cnet bandwidth meter, which downloads an image to your computer, mea- sures the time of the download, and compares it to the maximum speeds offered by other internet service pro- viders. there are several bandwidth meters available on the internet. when the network administrator checked the switches for network traffic, they showed low traffic, almost always less than percent of capacity. this was confusing: if the problem was neither with the bandwidth nor the vendors, what was causing the slow network performance? one of the university network administrators was consulted to see if any factor in their sphere could be having an effect on our network. we knew that the main university network had implemented a band- width shaper to regulate bandwidth. “these devices limit bandwidth . . . by greedy applications, guarantee mini- mum throughput for users, groups or protocols, and better utilize wide- area connections by smoothing out bursty traffic.” it was thought that perhaps this might be incorrectly pri- oritizing some of the library’s traffic. this was a dead end, though—the network administrators had stopped using the device. if the bandwidth was good and the traffic was manageable, then the problem appeared to not be at the library. however, according to bertot and mcclure, the bandwidth ques- tion is complex because typically an arbitrary number describes the number of kbps used to define “broadband.” . . . such arbi- trary definitions to describe bandwidth sufficiency are gen- erally not useful. the federal communications commission (fcc), for example, uses the term “high speed” for connections of kbps in at least one direc- tion. there are three problematic issues with this definition: . it specifies unidirectional bandwidth, meaning that a kbps download, but a much slower upload (e.g., kbps) would fit this defi- nition; . regardless of direction, bandwidth of kbps is neither high speed nor does it allow for a range of internet-based applications and services. this inad- equacy will increase sig- nificantly as internet-based applications continue to demand more bandwidth to operate properly. . the definition is in the con- text of broadband to the single user or household, and does not take into con- sideration the demands of a high-use multiple-worksta- tion public-access context. proxy issues auraria library uses the iii wam proxy server product. there were several things that pointed to the introducing zoomify image | smith playing tag in the dark: diagnosing slowness in library response time | brown-sica proxy being an issue. one was that the systems department had been experimenting with invoking the proxy in the library building in order to collect more accurate statistics and found that complaints about speed seemed to have started around the same time as this experiment. but if the bandwidth was not showing inadequacy and the traffic was light, why was this happening? the answer is better explained by azuma et al.: needless to say, busy web serv- ers must have many simultane- ous http sessions, and server throughput is degraded when effective resource management is not considered, even with large network capacity. web proxy servers must also accommodate a large number of tcp connec- tions, since they are usually pre- pared by isps (internet service providers) for their customers. furthermore, proxy servers must handle both upward tcp connec- tions (from proxy server to web servers) and downward tcp connections (from client hosts to proxy server). hence, the proxy server becomes a likely spot for bottlenecks to occur during web document transfers, even when the bandwidth of the network and web server performance are adequate. testing was done from on campus and off campus, with and without using the proxy server. the results showed that the connection was faster without the proxy. when testing was done from the health sciences library at the university of colorado with the same type of server and proxy, the response time was much faster. the difference between auraria library and the other library is that the com- munity auraria library serves (the community college of denver, metropolitan state college, and the university of colorado–denver) has a much larger user population who overwhelmingly use databases from home, therefore taxing the proxy server. the other library belonged to a smaller campus, but the hardware was the same. the proxy was imme- diately dropped for on-campus users, and that resulted in some response- time improvements. a conference call was set up with the proxy ven- dor to determine if improvements in response time might be attained by changing from a proxy server to ldap (lightweight directory access protocol) authentication. the response given was that although there might be other benefits, increased response time was not one of them. library network hardware it was evident that the biggest bottle- neck was the proxy, so the systems department decided to take a closer look at iii’s hardware. the switch that regulated traffic between the network and the server that houses our integrated library system, part of which is the proxy server, was discovered to have been set at “half- duplex.” half-duplex refers to the trans- mission of data in just one direc- tion at a time. for example, a walkie-talkie is a half-duplex device because only one party can talk at a time. in contrast, a telephone is a full-duplex device because both parties can talk simultaneously. duplex modes often are used in reference to network data transmissions. some modems contain a switch that lets you select between half- duplex and full-duplex modes. the correct choice depends on which program you are using to transmit data through the modem. when this setting was changed to full duplex response time increased. there was also concern that this switch had not been functioning as well as it could. the switch was replaced, and this also improved response time. in addition, the old server purchased through iii was a generic server that had specifi- cations based on the demands of the ils software and didn’t into consid- eration the amount of traffic going to the proxy server. auraria library, which serves a campus of more than thirty thousand full-time equivalent students, is a library with one of the largest commuter student popula- tions in the country. a new server had been scheduled to be purchased in the near future, so a call was made to the ils vendor to talk about our hypoth- esis and requirements. the vendor agreed that the library should change the specification on the new server to make sure it served the library’s unique demands. a server will be purchased with increased memory and a second processor to hopefully keep these problems from happening again in the next few years. also, the cabling between the switch and the server was changed to greater facili- tate heavy traffic. conclusion although it is sometimes a daunting task to try to discover where prob- lems occur in the library’s database response time because there are so many contributing factors and because librarians often do not feel that they have enough technical knowledge to analyze such problems, there are cer- tain things that can be examined and analyzed. it is important to look at how each library is unique and may be inadequately served by current band- width and hardware configurations. it is also important not to be intimidated by computer science literature and to trust patterns of reported problems. the auraria library systems depart- ment was fortunate to also be able to compare problems with colleagues at other libraries and test in those librar- ies, which revealed issues that were unique and therefore most likely due to a problem at the library end. it is important to keep learning about how information technology and libraries | december information technology and libraries | december your system functions and to try to diagnose the problem by slowly look- ing at one piece at a time. though no one ever seems to be completely satis- fied with the speed of their network, the employees of auraria library, especially those who work with the public, have been pleased with the increased speed they are experiencing when using proprietary databases. having improved on the response- time speed issue, other problems that are not caused by the proxy hard- ware have been illuminated, such as browser configuration, which may be hampering certain databases—some- thing that had been attributed to the network. references . webopedia, s.v. “bottleneck,” www.webopedia.com/term/b/bottle- neck.html (accessed oct. , ). . david r. gerhan and stephen mutula, “bandwidth bottlenecks at the university of botswana,” library hi tech , no. ( ): – . claudine badue et al., “basic issues on the processing of web queries,” sigir forum; proceedings (new york: asso- ciation for computing machinery, ): – . . john carlo bertot and charles r. mcclure,” assessing sufficiency and quality of bandwidth for public librar- ies,” information technology and librar- ies , no. (mar. ): – . . kazuhiro azuma, takuya oka- moto, go hasegawa, and murata mas- ayuki, “design, implementation and evaluation of resource management sys- tem for internet servers,” journal of high speed networks , no. ( ): – . . “cnet bandwidth meter,” http:// reviews.cnet.com/internet-speed-test (accessed oct. , ). . michael j. demaria, “warding off wan gridlock,” network computing nov. , , www.networkcomputing.com/ showitem.jhtml?docid= f (accessed oct. , ). . bertot and mcclure, “assessing sufficiency and quality of bandwidth for public libraries,” . . azuma, okamoto, hasegawa, and masayuki, “design, implementation and evaluation of resource management sys- tem for internet servers,” . . webopedia, s.v. “half-duplex,” www.webopedia.com/term/h/half _duplex.html (accessed oct. , ). lita cover , cover , cover index to advertisers information technology and libraries | june author id box for column layout column title information technology and libraries | june communications michaela brenner and peter klein discovering the library with google earth libraries need to provide attractive and exciting discovery tools to draw patrons to the valuable resources in their catalogs. the authors con- ducted a pilot project to explore the free version of google earth as such a discover tool for portland state library’s digital collection of urban planning documents. they created eye-catching placemarks with links to parts of this collection, as well as to other pertinent materials like books, images, and historical background information. the detailed how-to-do part of this article is preceded by a discussion about discovery of library materials and followed by possible applications of this google earth project. in calhoun’s report to the library of congress, it becomes clear that staff time and resources will need to move from cataloging traditional formats, like books, to cataloging unique primary sources, and then providing access to these sources from many different angles. “organize, digitize, expose unique special collections” (calhoun ). in , portland state university library received a grant “to develop a digital library under the sponsor- ship of the portland state university library to serve as a central repository of the collection, accession, and dis- semination of [urban] key planning documents . . . that have high value for oregon citizens and for scholars around the world” (abbott ). this collection is called the oregon sustainable community digital library (oscdl) and is an ongoing project that includes literature, plan- ning reports, maps, images, rlis (regional land information system) geographical data, and more. much of the older material is unpublished, and making it available online pres- ents a valuable resource. most of the digitized—and, more recently, born- digital—documents are accessible through the library’s catalog, where patrons can find them together with other library materials about the city of portland. the bibliographic records are arranged in the catalog in an electronic resource management (erm) system (brenner, larsen, and weston ). additionally, these bibliographic data are regularly exported from the library catalog to the oscdl web site (http://oscdl. research.pdx.edu) and there inte- grated with gis (global information system) features, thus optimizing cataloging costs by reusing data in a different electronic environment. committed to not falling into the trap that clifford lynch had in mind when he wrote, “i think there is a mental picture that many of us have that digitization is something you do and you finish . . . a finite, one-time process“ (lynch ), and agreeing with gatenby that “it doesn’t matter at all if a user finds our opac through the ‘back door ’“ (gatenby ), the authors looked into further using these existing data from the library catalog by making them accessible from a popular and appealing place on the internet, a place that users are more likely to visit than the library catalog. the free version of google earth, a virtual-globe program that can be installed on pcs, lent itself to experimenting. “google earth com- bines the power of google search with satellite imagery, maps, terrain and -d buildings to put the world’s geographic information at your fin- gertips” (http://earth.google.com). from there, the authors provide links to the digitized documents in the library catalog. easy distribution, as well as the more playful nature of this pilot project and the inclusion of pictures, make the available data even more attractive to users. “google now reigns” “google now reigns,” claims karen markey (markey ), and many others agree that using google is easier and more appealing to most than using library catalogs. google’s popularity has been growing spec- tacularly. in august , google accounted for percent of all u.s. searches (avtec media group ). in contrast, the oclc report on how users perceive the library shows that only one percent of the respondents begin their information search on a library web site, while percent use search engines (de rosa, et al. ). “if we [libraries] want to survive,” says stephen abram, “we must place our messages where the users are seeking answers and will trip over them. today that usually means at yahoo, msn, and google” (abram ). according to lorcan dempsey, in the longer run, traffic to the library catalog will come by linking from larger consolidated resources, like open worldcat and google scholar (dempsey ). dempsey also stressed that it becomes more and more significant to differentiate between discov- ery and location (dempsey a). initially, users want to discover; they want to find what interests them independent from where this information is actually located and available. while there may be lots of valuable, detailed, and exceptionally well-organized bibliographic infor- mation in the library catalog, not michaela brenner (brennerm@pdx.edu) is assistant professor and database maintenance and catalog librarian at portland state university library, oregon. peter klein (peter.klein@colorado.edu) is aerospace engineering bs/ms at the university of colorado at boulder. introducing zoomify image | smith discovering the library with google earth | brenner and klein many users (one percent) are willing to discover this information through the catalog. they may not discover what a library has to offer if “the library does not find a way to go to the user, rather than waiting for the user to come to the library” (coyle ). unless the intent is to keep our treasures buried, the library com- munity needs to work with popular outside discovery environments— like search engines—to bring infor- mation available in libraries to users from the outside. libraries are, although sometimes reluctantly, responding. google, google scholar, and google books are open worldcat partner sites that are now or soon will be providing access to worldcat records. google book search includes “find this book in the library,” and the advanced book search also has the option to limit a search to library catalogs with access to the worldcat web record for each item. “deep linking” enables web users to link from search results in yahoo, google, or other partner sites to the “find in a library” interface in open worldcat, and then directly to the item’s record in their library’s online public access catalog (opac). simply put, “find it on google, get it from your library” (calhoun ). the “leveraged discovery envi- ronment” is an expression coined by dempsey that means it becomes increasingly important to leverage a “discovery environment which is outside your control to bring peo- ple back into our catalog environ- ment (like amazon, google scholar)” (dempsey b). issues in calhoun’s report to the library of congress include the ques- tion of how to get a google user from google to library collections. she quotes an interviewee saying that “data about a library’s collec- tion needs to be on google and other popular sites as well as the library interface” (calhoun ). with evidence pointing to the heavy use of google for discovery and with google earth technology providing such a powerful visualiza- tion tool, the authors felt tempted to experiment with existing data from portland state library’s digital oscdl collection and make these data accessible through a virtual globe. the king’s college cultural heritage project martyn jessop from king’s college in london, united kingdom, published an article about a relatively small pilot project on providing access to a digital cultural heritage collection through a geographical informa- tion system (jessop ). jessop’s approach to explore different tech- nologies and techniques to apply to existing data about unique primary sources was exactly what the authors had in mind with this project, and provided encouragement to move forward with the idea of provid- ing additional access to the oregon sustainable community digital library (oscdl) collections through google earth. similar to jessop, the authors regard it an unaffordable lux- ury to put a great deal of effort into collecting, digitizing, and catalog- ing materials without making them available to a much broader audience through multiple access points. comparable to jessop, the goal of this project was to find a relatively simple, low-cost technological solu- tion that could also be applied to a much wider range of data without much more investment in staff time and money. once the authors mastered the ini- tial hurdle of understanding google earth’s programming language, they could easily identify with jessop’s notion of “project creep” as more and more possibilities arose to make the project more appealing. this, as with the king’s college project, was a valuable part of the develop- ment process, the details of which are described below. the portland state library oscdl-on- google-earth project the authors chose ten portland- based oscdl sub-collections as the basis of this pilot project: harbor drive, front street, portland public market, urban studies collection, downtown, park blocks, south park blocks, pioneer courthouse square, portland city archives, and jpact (joint policy advisory committee on transportation). the programming language for google earth is kml (keyhole markup language), a file format used to display geographic data. kml is based on the xml standard and can be created with the google earth user interface or from scratch with a simple text editor. having no pre- vious kml experience, the authors decided to use both. figure . basic placemark in google earth figure . kml script for basic placemark information technology and libraries | june information technology and libraries | june a basic placemark provided by google earth (figure ), copied and pasted in notepad (figure ), was the starting point. at portland state library, information technology routinely batch export cataloged oscdl data from the library catalog (ils) to the oscdl web site to reuse them. for the google earth project, the authors had two options, to either export data relevant to our collections from the ils to a spreadsheet or to use an existing excel spreadsheet contain- ing most of the same data, including place coordinates. this spreadsheet was one of many others that had been created to keep track for the digitization process as well as for creating bibliographic records for the library catalog later. using the avail- able spreadsheet again, the following data were retained: n the title of the collection n longitude and latitude of the place the collection refers to n a brief description of the collec- tion the following were added manu- ally to the remaining spreadsheet: n all the texts and urls for the collection-specific links n urls for the collection-specific images the authors extracted the place- mark-specific script from figure to create a template in notepad. a general description and all links that were the same for the ten collec- tions were added to this template, and placeholders were inserted for collection-specific data (figure ). using microsoft office word’s mail merge, the authors populated the template with the data from the spreadsheet in one quick step. the result was a kml script that included all the placemark data for the ten col- lections (figure ). the script was saved as plain text (.txt) first, and then renamed with the extension .kml, which represents the final file (figure ). clicking the oscdl.kml icon on a desktop or inside a web application opens google earth. the user “flies” to portland, where ten stars represent the ten collections (figure ). zooming in, the placemarks show the locations to which the collections refer. considering the many layers and icons available in google earth, the authors decided to use yellow stars to make them more visible. in order to avoid clutter and overlap- ping labels, titles only appear on mouse-over (figures and ). figure shows the open place- mark for portland public market. “portland state university” with the university’s logo is a link that takes the user to the university’s homepage. the next line is the title of the collection, followed by a brief description. the paragraph after that is the same for all collections and includes links to the portland state university library and the oscdl web site. the collection-specific links that follow next go to the library catalog where the user has access to the digitized manuscripts of this collection (figure ). other pertinent links—in this case to a book available in the library, a public web site on the history of the market, and a historic image of the market—were added as well. to make the placemarks visu- ally more attractive, all links are pre- sented in the school’s “psu green,” and an image representative of the collection was added. the pictures can be enlarged in a new window by clicking on them. to avoid copyright issues, the authors photographed their own images. the last link opens an e-mail window for questions and comments (figure ). this link is intended to bring some feedback and suggestions on how to improve the project and on its value for researchers and other users. the authors have been toying with the idea of including in the future more elaborate features such as video clips and music. one more recent feature is that kml files, created in google earth, can now also be viewed on the web by simply entering the url of the kml file into the search box of google maps (figure ), thus cre- ating google earth placemarks in figure . detail of template with variables between « double brackets » figure . detail: “downtown” placemark of finished kml script figure . simplified process figure . ten stars representing the ten collections introducing zoomify image | smith discovering the library with google earth | brenner and klein google maps with different view options (figures and ). not all formatting is correctly transferred, and at this point, there is no way to correct this in google maps. for example, the yellow stars were white, the mouse-over didn’t work and the size of the placemarks was impre- cise. however, the content of the placemarks—except for the images which didn’t show on some comput- ers—was fully retained and all links worked (figure ). although the use of the kml file in google maps is not as elegant as in google earth, it has the advantage that there is no need to install software as with google earth. this adds value to kml files and makes projects like this more versatile. the authors have identified sev- eral uses for the kml file: n a workstation in the library can be dedicated to resources about the city of portland. an icon on the desktop of this workstation will open google earth and “fly” directly to portland where the yellow stars are displayed. n professors can easily add the .kml file to webct (now blackboard) or other course management sys- tems. n the file can be e-mailed as an figure . zoomed in with mouse-over placemark figure . location of the pioneer courthouse square placemark figure . portland public market figure . access to the collection in library catalog figure . ready-to-go e-mail window figure . url of kml file in google maps search box figure . “map” view in google maps figure . “satellite” view in google maps figure . portland public market place- mark in google maps information technology and libraries | june information technology and libraries | june attachment to those interested in the development of the city of portland. n a link from the wikipedia page related to the oscdl project leads to the google earth pilot project. n the project was added to the google earth gallery where many remarkable projects, cre- ated by individuals and groups can be found. n it can also be accessed through the oscdl web site, and rele- vant links from the records in the library catalog to google maps can be included. it may be use- ful to alert patrons, who actually did come to the catalog by them- selves, to this visual tool. conclusion “the question now is not how we improve the catalog as such,” says dempsey. “it is how we provide effec- tive discovery and delivery of library materials in a network environment where attention is scarce and infor- mation resources are abundant and where discovery opportunities are being centralized into major search engines and distributed to other envi- ronments” (dempsey a). with this in mind, the authors took on the challenge to create another discovery tool for one of the library’s primary unique digital collections. google earth is not the web, and it needs to be installed on a worksta- tion in order to use a kml file. on the other hand, the file created in google earth can also be used on the web more readily but less elegantly in google maps, thus possibly reach- ing a larger audience. similar to the king’s college project and following abram’s sug- gestion that “we should experiment more with pilots in specific areas” (abram ), this pilot project is of an exploratory, experimental nature. and as with many experiments, the authors were testing an idea, trying something different and new to find out how useful this idea might be, and useful applications for this proj- ect were identified. google earth is a sophisticated, attractive, and exciting program—and fun to play with. in a time “where attention is scarce and information resources are abundant,” as dempsey ( a) says, we need to provide these kinds of discovery tools to attract patrons and to lure them to these valuable resources in our library’s catalog that we created with so much diligence and cost of staff time and resources. works cited abbott, carl. . planning a sustain- able portland: a digital library for local, regional, and state planning and policy documents. framing paper. http://oscdl.research.pdx.edu/docu- ments/library_grant.pdf. abram, stephen. . the google oppor- tunity. library journal , no. : . avtec media group. . search engine statistics. http://avtecmedia.com/ internet-marketing/internet-market- ing-trends.htm. brenner, michaela, tom larsen, and clau- dia weston. . digital collection management through the library cata- log. information technology and libraries , no. : – . calhoun, karen. . the changing nature of the catalog and its integra- tion with other discovery tools; final report, prepared for the library of congress. www.loc.gov.proxy.lib.pdx. edu/catdir/calhoun-report-final.pdf. coyle, karen. . the library catalog in a . world. the journal of academic librarianship , no. : – . de rosa, cathy et al. . perceptions of libraries and information resources. a report to the oclc membership. www .oclc.org.proxy.lib.pdx.edu/reports/ pdfs/percept_all.pdf. dempsey, lorcan. a. the library catalogue in the new discovery envi- ronment: some thoughts. ariadne . www.ariadne.ac.uk/issue /demp- sey. dempsey, lorcan. b. lifting out the catalog discovery expe- rience. lorcan dempsey’s weblog on libraries, services, and networks, may , . http://orweblog .oclc.org/archives/ .html dempsey, lorcan. . making data work—web . and catalogs. lor- can dempsey’s weblog on libra- ries, services, and networks, october , . http://orweblog.oclc .org/archives/ .html gatenby, janifer. . accessing library materials via google and other web sites. paper presented to elag (euro- pean library automation group), may , . http://elag .upf. edu/papers/gatenby_ .pdf. jessop, martyn. . the application of a geographical information system to the creation of a cultural heritage digital resource. literary and linguistic computing: journal of the association for literary and linguistic computing , no. : – . lynch, clifford. . digital collections, digital libraries, and the digitization of cultural heritage information. first monday , no. . www.firstmonday. org/issues/issue _ /lynch. markey, karen. . the online library catalog. d-lib magazine , no. / . www .dlib.org/dlib/january /markey/ markey.html. lita cover , cover , cover index to advertisers information technology and libraries | december thomas sommer unlv special collections in the twenty-first century university of nevada las vegas (unlv) special collections is consistently striving to provide several avenues of dis- covery to its diverse range of patrons. specifically, unlv special collections has planned and implemented several online tools to facilitate unearthing treasures in the collec- tions. these online tools incorporate web . features as well as searchable interfaces to collections. t he university of nevada las vegas (unlv) special collections has been working toward creating a visible archival space in the twenty-first century that assists its patrons’ quest for historical discovery in unlv’s unique southern nevada, gaming, and las vegas collections. this effort has helped patrons ranging from researchers to students to residents. special collections has created a discovery environment that incorporates several points of access, including virtual exhibits, a collection-wide search box, and digital collections. unlv special collections also has added web . features to aid in the discovery and enrichment of this historical infor- mation. these new features range from a what’s new blog to a digital collection with interactive features. the first point of discovery within the unlv special collections website began with the virtual exhibits. staff created the virtual exhibits as static html pages that showcased unique materials housed within unlv special collections. they showed the scope and diversity of materials on a specific topic available to researchers, faculty, and students. one virtual exhibit is “dino at the sands” (figure ), a point of discovery for the history not only of dean martin but of many rat pack exploits. the photographs in this exhibit come from the sands collection. it is a static html page, and it provides information and pictures regarding one of las vegas’ most famous entertainers. this exhibit contains links to rat pack information and various resources on dean martin, including photo- graphs, books, and videotapes. a second mode of discovery within the unlv special collections website is its new “search special collections” google-like search box (figure ). this is located on the homepage and searches the manuscript, photograph, and oral history primary source collections. the purpose is to aid in the discovery of material within the collections that is not yet detailed in the public online catalog. in the past researchers would have to work through the special collection’s website to locate the resources. they can now go to one place to search for various types of material—a one-stop shop. the search results are easy to read and highlight the search term (see figure ). the third point of access is the digital collection. these collections are digital copies of original materials located within the archives. the digital copies are presented online, described, and organized for easy access. each collection offers full-text searches, browsing, zoom, pan, figure . unlv special collections search box figure . “dino at the sands” exhibit thomas sommer (thomas.sommer@unlv.edu) is university and technical services archivist in special collections at the university of nevada las vegas libraries. unlv special collections in the twenty-first century | sommer side-by-side comparison, and exporting for presentation and reuse. the newest example of a digital collection is “southern nevada: the boomtown years” (figure ). this collection brings together a wide range of original materials from var- ious collections located within unlv special collections, the nevada state museum, the historical society in las vegas, and the clark county heritage museum. it even provides standards-based activities for elementary and high school students. this project was funded by the nevada state library and archives under the library services and technology act (lsta) as amended through the institute of museum figure . “southern nevada: the boomtown years” digital collection figure . “what’s new” blog figure . unlv special collection facebook page figure . hoover dam search results information technology and libraries | december and library services (imls). unlv special collections director peter michel selected the content. the team included fourteen members, four of whom were funded by the grant. christy keeler, phd, created the educator pages and designed the student activities. new collections are great, but users have to know they exist. to announce new collections and displays, special collections first added a what’s new blog that includes an rss feed to keep patrons up-to-date on new messages (figure ). another avenue of interaction was implemented in april when special collections created its own facebook page (figure ). students and researchers are encouraged to become fans. status updates with images and links to southern nevada and las vegas resources lead the fans back to the main web- site where the other treasures can be discovered. special collections has implemented various web . features within its newest digital collections. specifically, it added a comments section, a “rate it” feature, and an rss feature to its latest digital collections (figures , , and ). these latest trends enrich the collections’ resources with patron-supplied information. as is apparent, unlv special collections imple- mented several online tools to allow patrons to discover its extensive primary resources. these tools range from virtual exhibits and digital collections with web . features to blogs and social networking sites. special collections has endeavored to stay on top of the latest trends to benefit its patrons and facilitate their discovery of historical materials in the twenty-first century. figure . “rate it” feature for aerial view of hughes aircraft plant photograph figure . comments section for aerial view of hughes aircraft plant photograph figure . rss feature for the index to the “welcome home howard” digital collection continued on page information technology and libraries | december as previously mentioned, these easy-to-use tools can allow screencast videos and screenshots to be integrated into a variety of online spaces. a particularly effective type of online space for potential integration of such screencast videos and screenshots are library “how do i find . . .” research help guides. many of these “how do i find . . .” research help guides serve as pathfinders for patrons, outlining processes for obtaining information sources. currently, many of these pathfinders are in text form, and experimentation with the tools outlined in this article can empower library staff to enhance their own pathfinders with screencast videos and screenshot tutorials. reference . “unlv libraries strategic plan – ,” http://www .library.unlv.edu/about/strategic_plan - .pdf (accessed july , ): . unlv special collections continued from page references . peter michel, “dino at the sands,” unlv special collec- tions, http://www.library.unlv.edu/speccol/dino/index.html (accessed july , ). . peter michel, “unlv special collections search box.” unlv special collections. http://www.library.unlv.edu/speccol/ index.html (accessed july , ). . unlv special collections search results, “hoover dam,” http://www.library.unlv.edu/speccol/databases/index .php?search_query=hoover+dam&bts=search&cols[]=oh&cols []=man&cols[]=photocoll&act= (accessed october , ). . unlv libraries, “southern nevada: the boomtown years,” http://digital.library.unlv.edu/boomtown/ (accessed july , ). . unlv special collections, “what’s new in special col- lections,” http://blogs.library.unlv.edu/whats_new_in_special_ collections/ (accessed july , ). . unlv special collections, “unlv special collections facebook homepage,” http://www.facebook.com/home .php?#/pages/las-vegas-nv/unlv-special-collections/ ?ref=search (accessed july , ). . unlv libraries, “comments section for the aerial view of hughes aircraft plant photograph,” http://digital.library .unlv.edu/hughes/dm.php/hughes/ (accessed july , ); unlv libraries, “‘rate it’ feature for the aerial view of hughes aircraft plant photograph,” http://digital.library.unlv.edu/ hughes/dm.php/hughes/ (accessed july , ); unlv libraries, “rss feature for the index to the welcome home how- ard digital collection” http://digital.library.unlv.edu/hughes/ dm.php/ (accessed july , ). statement of ownership, management, and circulation information technology and libraries, publication no. - , is published quarterly in march, june, september, and december by the library information and technology association, american library association, e. huron st., chicago, illinois - . editor: marc truitt, associate director, information technology resources and services, university of alberta, k adams/cameron library and services, university of alberta, edmonton, ab t g j canada. annual subscription price, $ . printed in u.s.a. with periodical-class postage paid at chicago, illinois, and other locations. as a nonprofit organization authorized to mail at special rates (dmm section . only), the purpose, function, and nonprofit status for federal income tax purposes have not changed during the preceding twelve months. extent and nature of circulation (average figures denote the average number of copies printed each issue during the preceding twelve months; actual figures denote actual number of copies of single issue published nearest to filing date: september issue). total number of copies printed: average, , ; actual, , . mailed outside country paid subscriptions: average, , ; actual, , . sales through dealers and carriers, street vendors, and counter sales: average, ; actual . total paid distribution: average, , ; actual, , . free or nominal rate copies mailed at other classes through the usps: average, ; actual, . free distribution outside the mail (total): average, ; actual, . total free or nominal rate distribution: average, ; actual, . total distribution: average, , ; actual, , . office use, leftover, unaccounted, spoiled after printing: average, ; actual, . total: average, , ; actual, , . percentage paid: average, . ; actual, . . s t a t e m e n t o f o w n e r s h i p , m a n a g e m e n t , a n d c i r c u l a t i o n ( p s f o r m , s e p t e m b e r ) f i l e d w i t h t h e u n i t e d s t a t e s p o s t o f f i c e p o s t m a s t e r i n c h i c a g o , o c t o b e r , . fboze rectangle editorial | truitt marc truitteditorial: computing in the “cloud” silver lining or stormy weather ahead? c loud computing. remote hosting. software as a service (saas). outsourcing. terms that all describe various parts of the same it elephant these days. the sexy ones—cloud computing, for example—empha- size new age-y, “ . ” virtues of collaboration and sharing with perhaps slightly mystic overtones: exactly where and what is the “cloud,” after all? others, such as the more utilitarian “remote hosting” and “outsourcing,” appeal more to the bean counters and sustainability- minded among us. but they’re really all about the same thing: the tradeoff between cost and control. that the issue increasingly resonates with it opera- tions at all levels these days can be seen in various ways. i’ll cite just a few: n at the meeting of the lita heads of library technology (holt) interest group at the ala annual conference in chicago, two topics dominated the list of proposed holt programs for the annual conference. one of these was the question of virtualization technology, and the other was the whole white hat–black hat dichotomy of the cloud. practically everyone in the room seemed to be looking at—or wanting to know more about—the cloud and how it might be used to ben- efit institutions. n my institution is considering outsourcing e-mail. all of it—to google. times are tough, and we’re being told that by handing e-mail over to the googleplex, our hardware, licensing, evergreen- ing, and technical support fees will total zero. zilch. with no advertising. heady stuff when your campus hosts thirty-plus central and depart- mental mail servers, at least as many blackberry servers, and total costs in people, hardware, licensing, and infrastructure are estimated to exceed can$ , , annually. n in the last couple of days, library electronic dis- cussion lists such as web lib have been abuzz— or do we now say a-twitter?—about amazon’s orwellian kindle episode, in which the firm deleted copies of and animal farm from subscribers’ kindle e-book readers without their knowledge or consent. indeed, amazon’s action was in violation of its own terms of service, in which the company “grants [the kindle owner] the non-exclusive right to keep a permanent copy of the applicable digital content and to view, use, and display such digital content an unlim- ited number of times, solely on the device or as authorized by amazon as part of the service and solely for [the kindle owner ’s] personal, non- commercial use.” all of this has me thinking back to the late s marketing slogan of a manufacturer of consumer-grade mass storage devices—remember removable hard drives? iomega launched its advertising campaign for the gb jaz drive with the catch-line “because it’s your stuff.” ultimately, whether we park it locally or send it to the cloud, i think we need to remember that it is our stuff. what i fear is that in straitened times, it becomes easy to forget this as we struggle to balance limited staff, infra- structure, and budgets. we wonder how we’ll find the time and resources to do all the sexy and forward-looking things, burdened as we are with the demands of support- ing legacy applications, “utility” services, and a huge and constantly growing pile of all kinds of content that must be stored, served up, backed up (and, we hope, not too often, restored), migrated, and preserved. the buzz over the cloud and all its variants thus has a certain siren-like quality about it. the notion of sign- ing over to someone else’s care—for little or no apparent cost—our basic services and even our own content (our stuff) is very appealing. the song is all the more persua- sive in a climate where we’ve moved from just the normal bad news of merely doing more with less to a situation where staff layoffs are no longer limited to corporate and public libraries, but indeed extend now to our greatest institutions. at the risk of sounding like a paranoid naysayer to what might seem a no-brainer proposition, i’d like to sug- gest a few test questions for evaluating whether, how, and when we send our stuff into the cloud: . why are we doing this? what do we hope to gain? . what will it cost us? bear in mind that nothing is free—except, in the open-source community, where free beer is, unlike kittens, free. if, for example, the borg offer to provide institutional mail without advertisements, there is surely a cost somewhere. the borg, sensibly enough, are not in business to provide us with pro bono services. . what is the gain or loss to our staff and patrons in terms of local customization options, functionality, access, etc? . how much control do we have over the ser- vice offered or how our content is used, stored, marc truitt (marc.truitt@ualberta.ca) is associate university librarian, bibliographic and information technology services, university of alberta libraries, edmonton, alberta, canada, and editor of ital. information technology and libraries | september repurposed, or made available to other parties? . what’s the exit strategy? what if we want to pick up and move elsewhere? can we reclaim all of our stuff easily and portably, leaving no sign that we’d ever sent it to the cloud? we are responsible for the services we provide and for the content we have been entrusted. we cannot shrug off this duty by simply consigning our services and our stuff to the cloud. to do so leaves us vulnerable to an irreparable loss of credibility with our users; eventually some among them would rightly ask, “so what is it that you folks do, anyway?” we’re responsible for it—whether it’s at home or in the cloud—because it’s our stuff. it is our stuff, right? references and notes . i should confess, in the interest of full disclosure, that it was eli neiburger of the ann arbor district library who suggested “hosted services as savior or slippery slope” for next year’s holt program. i’ve shamelessly filched eli’s topic, if not his catchy title, for this column. thanks, eli. also, again in the interest of full disclosure, i suggested the virtualization topic, which eventually won the support of the group. finally, some participants in the discussion observed that virtualization technology and hosting are in many ways two sides of the same topical coin, but i’ll leave that for others to debate. . brad stone, “amazon erases orwell books from kin- dle,” new york times, july , , http://www.nytimes .com/ / / /technology/companies/ amazon.html?_ r= (accessed july , ). . amazon.com, “amazon kindle: license agreement and terms of use,” http://www.amazon.com/gp/help/customer/ display.html?nodeid= (accessed july , ). . “budget cutbacks announced in libraries, center for pro- fessional development,” stanford university news, june , , http://news.stanford.edu/news/ /june /layoffs- .html (accessed july , ; “harvard libraries cuts jobs, hours,” harvard crimson (online edition), june, , http:// www.thecrimson.com/article.aspx?ref= (accessed july , ). catqc and shelf-ready material | jay, simpson, and smith michael jay ([e-mail?]) is information technology expert, software unit, information technology department; betsy simpson is chair, cataloging and metadata department; and doug smith is head, copy cataloging unit, cataloging and metadata department, george a. smathers libraries, university of florida, gainesville. michael jay, betsy simpson, and doug smith catqc and shelf-ready material: speeding collections to users while preserving data quality libraries contract with vendors to provide shelf-ready material, but is it really shelf-ready? it arrives with all the physical processing needed for immediate shelving, then lingers in back offices while staff conduct item- by-item checks against the catalog. catqc, a console application for microsoft windows developed at the university of florida, builds on oclc services to get material to the shelves and into the hands of users with- out delay and without sacrificing data quality. using standard c programming, catqc identifies problems in marc record files, often applying complex condition- als, and generates easy-to-use reports that do not require manual item review. a primary goal behind improvements in technical service workflows is to serve users more efficiently. however, the push to move material through the system faster can result in shortcuts that undermine bib- liographic quality. developing safeguards that maintain sufficiently high standards but don’t sacrifice productiv- ity is the modus operandi for technical service managers. the implementation of oclc’s worldcat cataloging partners (wcp, formerly promptcat) and bibliographic record notification services offers an opportunity to retool workflows to take advantage of automated pro- cesses to the fullest extent possible, but also requires some backroom creativity to assure that adequate access to material is not diminished. n literature review quality control has traditionally been viewed as a central aspect of cataloging operations, either as part of item-by- item handling or manual and automated authority main- tenance. how this activity has been applied to outsourced cataloging was the subject of a survey of academic librar- ies in the united states and canada. a total of percent of libraries in the survey indicated that they forgo quality control of outsourced copy, primarily for government documents records. however, most respondents reported they review records for errors. of that group, percent focus on access points, percent check a variety of fields, and a significant minority— percent—look at all data points. overall, the libraries expressed satisfaction with the outsourced cataloging using the following measures of quality supplied by the author: accuracy, consistency, adequacy of access points, and timeliness. at the incep- tion of oclc’s promptcat service in , ohio state university libraries participated in a study to test similar quality control criteria with the stated goals of improving efficiency and reducing copyediting. the results were so favorable that the author speculated that promptcat would herald a future where libraries can “reassess their local practices and develop greater confidence in national standards so that catalog records can be integrated into local opacs with minimal revision and library hold- ings can be made available in bibliographic databases as quickly as possible.” fast forward a few years and the new incarnation of promptcat, wcp, is well on its way to fulfilling this dream. in a recent investigation conducted at the university of arkansas libraries, researchers concluded that error review of copy supplied through promptcat is necessary, but the error rate does not warrant discontinuance of the service. the benefits in terms of time savings far outweigh the effort expended to correct errors, particularly when the focus of the review is to correct errors critical to user access. while the researchers examined a wide variety of errors, a primary consideration was series headings, particularly given the problems cited in previous studies and noted in the article. with the announcement by the library of congress (lc) to curtail its practice of providing controlled series access, the cataloging community voiced great con- cern about the effect of that decision on user access. the arkansas study determined that “the significant number of series issues overall (even before lc stopped perform- ing series authority work) more than justifies our concern about providing series authority control for the shelf-ready titles.” approximately one third of the outsourced copy across the three record samples studied had a series, and, of that group, percent needed attention, predominantly taking the form of authority record creation with associ- ated analysis and classification decisions. the overwhelming consensus among catalogers is that error review is essential. as far as can be determined, an underlying premise behind such efforts seems to be that it is done with the book in hand. but could there be a way to satisfy the concerns without the book in hand? certainly, validation tools embedded in library management sys- tems provide protections whether records are manually entered or batchloaded, and outsourced authority main- tenance services (for those who can use them) offer fur- ther control. but a customizable tool that allows libraries to target specific needs, both standards-based and local, without relying on item-by-item handling can contribute michael jay (emjay@ufl.edu) is information technology expert, software unit, information technology department; betsy simpson (betsys@uflib.ufl.edu) is chair, cataloging and metadata department; and doug smith (dougsmith@uflib.ufl .edu) is head, copy cataloging unit, cataloging and metadata department, george a. smathers libraries, university of florida, gainesville. information technology and libraries | march to an economy of scale demanded by an environment with shrinking budgets and staff to devote to manual bibliographic scrutiny. if that tool is viewed as part of a workflow stream involving local error detection at the receiving location as well as enhancement at the network level (i.e., oclc’s bibliographic record notification service), then it becomes an important step in freeing catalogers to turn their attention to other priorities, such as digitized and hidden collections. n local setting and workflow the george a. smathers libraries at the university of florida encompasses six branches that address the infor- mation needs of a diverse academic research campus with close to fifty thousand undergraduate and gradu- ate students. the technical services division, which includes the acquisitions and licensing department and the cataloging and metadata department, acquires and catalogs approximately forty thousand items annu- ally. seeking ways to minimize the handling of incoming material, beginning in the departments developed a workflow that made it possible to send shelf-ready incoming material directly to the branches after check-in against the invoice. shelf-ready items represent approxi- mately percent of the libraries’ purchased mono- graphic resources at this time. by using wcp record loads along with vendor-supplied shelf-ready process- ing, the time from receipt to shelf has been reduced significantly because it is no longer necessary to send the bulk of the shipments to cataloging and metadata. exceptions to this practice include specific categories of material that require individual inspection. the vendor is asked to include a flag in books that fall into many of these categories: n any nonprocessed book or book without a spine label n books with spine labels that have numbering after the date (e.g., vol. , no. ) n books with cds or other formats included n books with loose maps n atlases n spiral-bound books n books that have the words “annual,” “biennial,” or a numeric year in the title (these may be a serial add to an existing record or part of a series that will be established during cataloging) to facilitate a post–receipt record review for those items not sent to cataloging and metadata, acquisitions and licensing runs a local programming tool, catqc, which reports records containing attributes cataloging and metadata has determined necessitate closer exami- nation. figure is an example of the reports generated, which are viewed using the mozilla firefox browser. copy catalogers rotate responsibility for checking the report and revising records when necessary. retrieval of the physical piece is only necessary in the percent of cases where the item needs to be relabeled. n catqc report catqc analyzes the content of the wcp record file and identifies records with particular bibliographic coding, which are used to detect potential problems: . encoding levels , , , , e, j, k, m . with non-english subfield b . fields with subfields h, n, or p . fields with subfields a or b that contain numerals . fields with subfields a or b that contain red flag keywords . fields . fields with first indicator . fields without subfield . xx fields with second indicators , , , and the numbers following each problem listed below indicate which codes are used to signal the presence of a potential problem. minimal-level copy ( ) the library’s wcp profiles, currently in place for three vendors, are set up to accept all oclc encoding levels. with such a wide-open plan, it is important to catch records with minimal-level copy to assure that appro- priate access points exist and are coded correctly. the library encounters these less-than-full encoding levels infrequently. parallel records ( ) catqc identifies foreign library records that are candi- dates for parallel record treatment by indicating in the report if the has a non-english subfield b. the report includes a field if present to alert catalogers that a parallel record is available. volume sets ( , , ) the library does not generally analyze the individual volumes of multipart monographic sets (i.e., volume sets) even when the volumes have distinctive titles. these catcq and shelf-ready material | jay, simpson, and smith “volume,” “part,” and “number” as well as common abbreviations of those words (e.g., v. or vol.). serial vs. monograph treatment ( , ) titles owned by the library and classi- fied as serials sometimes are ordered inadvertently as monographs, result- ing in the delivery of a monographic record. a similar problem also occasion- ally arises with new titles. by detecting numerals, keywords, or the presence of one or more of the subfields in the field, we can quickly scan a list of records with these characteristics. of course, most of the records detected by catqc are false hits because of the broad scope of the search; however, it takes only a few minutes to scan through the record list. non-print formats ( ) the library does not receive records for any format other than print through wcp. consequently, detecting the presence of a subfield h in the field is a good signal that there may be a problem with the record. alternate titles ( ) alternate titles can be an important access point for library users. sometimes text that should properly be in subfield i (e.g., “at head of title”) of the field is placed in subfield a in front of the alternate title. this adversely affects user access to the title through browse searching. catqc checks for and reports the presence of a field. the cataloger can then quickly confirm that it is coded correctly. untraced series ( ) as a program for cooperative cataloging (pcc) partici- pant, the library opted to follow pcc practice to continue to trace series despite lc’s decision in to treat as untraced all series statements in newly cataloged records. because some libraries chose to follow lc in its decision, there has been an overall increase in the use of untraced series statements across all types of record-encoding volumes are added to the collection under the title of the set. the june decision by lc to produce individual volume records when a distinctive title exists caused concern about the integrity of the libraries’ existing open volume set records. because such records typically have enumeration indicated in the subfield n, and sometimes p, of the field, the program searches for instances of those subfields. in addition, the program detects the presence of numerals in the and keywords such as figure . an example report from catcq information technology and libraries | march levels. to address this issue, catqc searches all wcp records for fields with first indicator . catalogers check the authority files for the series and make any necessary changes to the records. this is by far the most frequent correction made by catalogers. links ( ) to provide users with information about the nature of the urls displayed in the catalog, catalogers insure that explanatory text is recorded in subfield of the field. catqc looks for the absence of subfield , and, if absent, displays the field in the report as a hyperlink. the cata- loger adds the appropriate text (e.g., full text) as needed. subject headings with second indicators , , , and ( ) the catqc report reviewed by catalogers includes sub- ject headings with second indicator . when these head- ings duplicate headings already on the record, catalogers delete them from our local system. when the headings are not duplicates, the catalogers change the second indi- cator to . typically, xx fields with second indicators , , and contain non-english headings based on foreign thesauri. these headings can conflict with lc headings and, in some cases, are cross references on lc authorities. the resulting split files are not only confusing to patrons, but also add to the numbers of errors reported that require authority maintenance. for these reasons, our policy is to delete the headings from our local system. catqc detects the presence of second indicators , , or and creates a modified file with the headings removed with one excep- tion: a heading with second indicator and subfield of “nasat,” which indicates the heading is taken from the national aeronautics and space administration the- saurus, is not removed because the local preference is to retain the “nasat” headings. n library-specific issues catqc resolves local problems when needed. for exam- ple, when more than one lc call number was present on the record, the wcp spine manifest sent to the ven- dor used to contain the second call number, which was affixed to the item. when the wcp records were loaded into the library’s catalog, the first call number populated the holding. as a result, there was a discrepancy between the spine label on the book and the call number in the catalog. prior to generating the report, catqc found mul- tiple instances of call numbers in the records in the wcp file and created a modified file with the call numbers reordered so that the correct call number was used on the holding when the record was loaded. previously, the library’s opac did not display the text in subfield of the field, which specifies the type of material covered by the link, and to the user it appeared that the link was to a full-text resource. this was par- ticularly troublesome for records with lc links to table of contents, publisher descriptions, contributor information, and sample text. to prevent user frustration, catqc was programmed to move the links on the wcp records to xx fields. when the opac interface improved and the pro- gramming was no longer necessary, catqc was revised. n analysis to see how well catqc and oclc’s bibliographic notification service were meeting our goal of maintain- ing high-quality bibliographic control, reports were randomly selected from the reports generated by catqc between october and april . catqc found no problems in twelve ( percent) of the selected reports. these twelve were not used in the analysis, leav- ing fifty-one catqc reports examined with at least one potential problem flagged for review. an average of . percent of the records in the sample of reports was flagged as requiring review by a cataloger. an average of thirteen possible problems was detected per report. of these, percent were potential problems requiring at least some attention from the cataloger. the action required of the cataloger varied from simply check- ing the text of a field displayed in the report (e.g., fields) to bringing up the record in aleph and editing the bibliographic record (e.g., verifying and correcting series headings or eliminating unwanted subject headings). why the relatively high rate of false positives ( per- cent)? to minimize missing serials and volumes belong- ing to sets, catqc is designed to err on the side of caution. two of the criteria listed earlier were responsible for the vast majority of the false positives generated by catqc: fields with subfields a or b that contain numerals and fields with subfields a or b that contain red-flag keywords. clearly, if every record with a numeral in the is flagged, a lot of hits will be generated that are not actual problems. the list of keywords was purposefully designed to be extensive. for example, “volume,” “vol.,” and “v.” are all triggers causing a record to be flagged. therefore a bibliographic record containing the phrase “volume cost profit analysis” in the field would be flagged as a potential problem. at first glance, a report filled with so many false posi- tives may seem inefficient and burdensome for catalogers to use; however, this is largely mitigated by the excellent display format. the programmer worked closely with catcq and shelf-ready material | jay, simpson, and smith the copy cataloging unit staff to develop a user-friendly report format. each record is framed separately, making it easy to distinguish from adjoining records. potential problems are highlighted with red lettering immediately alerting catalogers to what the potential problem might be. whenever a potential problem is found, the text of the entire field appears in the report so that catalogers can see quickly whether the field triggering the flag is an actual problem. it takes a matter of seconds to glance through the fields of half a dozen records to see if the numeral or keyword detected is a problem. the catalogers who work with these reports estimated that it took them between two and three hours per month to both review the files and make corrections to bibliographic records. a second component of bibliographic quality main- tenance is oclc’s bibliographic record notification service. this service compares newly upgraded oclc records with records held by the library and delivers the upgraded records to the library. because catqc flags records with encoding levels of , , , , e, j, k, and m, it was possible to determine if these records had, in fact, been upgraded in oclc. in the sample, thirty-three records were flagged because of the encoding level. no upgrade had been made to . percent of the records in oclc as of august . upgrades had been made to . percent of the records. the remaining . percent of the records were manu- ally loaded by catalogers in copy cataloging. these typi- cally are records for items brought to copy cataloging by acquisitions and licensing because they meet one or more of the criteria for individual inspection discussed previously. when catalogers search oclc and find that the received record has not been upgraded, they search for another matching record. a third of the time, a record of higher quality than that received is found in oclc and exported to the catalog. the reason why the record of better quality is not harvested initially is not clear. it is possible that at the time the records were harvested both records were of equivalent quality and by chance one was enhanced over another. in no instance had any of the records originally harvested been upgraded (this is not reflected in the . percent of records not upgraded). encoding level records are excluded from catqc reports. because of the relatively quick turnaround for upgrades of this type of copy, the library decided to rely solely on the bibliographic record notification service. n technical specifications catqc is a console application for windows. written in standard c, it is designed to be portable to multiple oper- ating systems with little modification. no graphic inter- face was developed because (a) the users are satisfied with the current operating procedure and (b) the treat- ment of the records is predefined as a matter of local policy. the user opens a command console (cmd.exe) and types “catqc”+space+“[name of marc file]”+enter. the corrected file is generated; catqc analyzes the modified file and creates the xml report. it moves the report to a reviewing folder on a file server across the lan and indicates to the user that it is terminating. modifications require action by a programmer; the user cannot choose from a list of options. benefits include a kb file size and a processing speed of approximately , records per second. no quantitative analysis has yet been done related to the speed of processing, but to the user the entire process seems nearly instantaneous. the genesis of the project was an interest in the record structure of marc files brought about in the program- mer by the use of earlier local automation tools. the proj- ect was speculative. the first experiment contained the programming structure that would become catqc. one record is read into memory at a time, and there is another array held for individual marc fields. conceptually, the records are divided into three portions—leader, directory, and dataset—when the need arises to build an edited record. initially there was no editing, only the production of the report. the generation of strict, valid xml is a significant aspect of catqc. an original document type was created, along with a corresponding cascading style sheet. the reports are viewable to anyone with an xml–capable browser either through file server, web server, or e-mail. (the current version of internet explorer does not fully support the style sheet syntax.) this continues to be con- venient for the report reviewers because they do not have to be client application operators. see appendix a for an excerpt of a document instance and appendix b for the document type definition. catqc is not currently a generalized tool such as marcedit, a widely used marc editing utility that pro- vides a standard array of basic capabilities: field count- ing, field and subfield deletion (with certain conditional checks), field and subfield additions, field swapping and text replacement, and file conversion to and from vari- ous formats such as marcxml and dublin core as well as between marc- and utf- encodings. marcedit continues to grow and does offer programmability that relies on the windows scripting host. this requires the user to either learn vbscript or use the wizards offered by marcedit. the catqc development goal was to create a report, viewable through a lan or the internet, which alerts a group of catalogers to potential problems with spe- cific records, often illustrating those problems. although it might have been possible to use a combination of marcedit capabilities and local programming to help achieve this goal, it likely would have been a more cumbersome route, particularly taking into consideration the multidimensional information technology and libraries | march conditionals desired. it was deemed easier to write a pro- gram that addresses local needs directly in a language already familiar to the programmer. as catqc evolved, it was modified to identify more potential problems and to do more logical comparisons as well as to edit the files as necessary before generating the reports. catqc addresses a particular workflow directly and provides one solution. it is procedural as opposed to event driven or object oriented. with version . , the generic functions were extracted into a marclib . , a common object file format library. functions specific to local workflow remain in catqc. the program is freely available to interested libraries by contacting the authors. as of this writing, the university of florida plans to dis- tribute this utility under the gnu public license version (see www.opensource.org/licenses/gpl- . .html) while retaining copyright. n conclusion catqc provides catalogers an easy way to check the bibliographic quality of shelf-ready material without the book in hand. as a result, throughput time from receipt to shelf is reduced, and staff can focus data review on problem areas—those affecting access or interfering with local processes. some of the issues addressed by catqc are of concern to all libraries while others reflect local preferences. the program could be easily modified to conform to those preferences. automation tools such as catqc are of key importance to libraries seeking ways to streamline workflows to the benefit of users. references and notes . vinh-the lam, “quality control issues in outsourcing cataloging in united states and canadian academic libraries,” cataloging & classification quarterly , no. ( ): – . . mary m. rider, “promptcat: a projected service for automatic cataloging—results of a study at the ohio state university libraries,” cataloging & classification quarterly , no. ( ): . . mary walker and deb kulczak, “shelf-ready books using promptcat and ybp: issues to consider (an analysis of errors at the university of arkansas),” library collections, acquisitions, & technical services , no. ( ): – . . “lc pulls plug on series authority records,” cataloging & classification quarterly , no. ( ): – . . walker and kulczak, “shelf-ready books.” . for more information about marcedit, see http://oregon state.edu/~reeset/marcedit/html/index.php. wcp file analysis: records analyzed. record: oclc number: timestamp: . : |a difference algebra /|c levin alexander. h n p numerals keywords appendix a. catqc document instance excerpt catcq and shelf-ready material | jay, simpson, and smith
: |a algebras and applications ;|v v.
. . .
appendix b. catqc document type definition information technology and libraries | march information technology and libraries | june tutorial andrew darby and ron gilmour adding delicious data to your library website social bookmarking services such as delicious offer a simple way of devel- oping lists of library resources. this paper outlines various methods of incorporating data from a delicious account into a webpage. we begin with a description of delicious linkrolls and tagrolls, the simplest but least flexible method of dis- playing delicious results. we then describe three more advanced meth- ods of manipulating delicious data using rss, json, and xml. code samples using php and javascript are provided. o ne of the primary components of web . is social bookmark- ing. social bookmarking ser- vices allow users to store bookmarks on the web where they are avail- able from any computer and to share these bookmarks with other users. even better, these bookmarks can be annotated and tagged to provide multiple points of subject access. social bookmarking services have become popular with librarians as a means of quickly assembling lists of resources. since anything with a url can become a bookmark, such lists can combine diverse resource types such as webpages, scholarly articles, and library catalog records. it is often desirable for the data stored in a social bookmarking account to be displayed in the context of a library webpage. this creates consistent branding and a more professional appearance. delicious (http://delicious .com/), one of the most popular social bookmarking tools, allows users to extract data from their accounts and to display this data on their own websites. delicious offers mul- tiple ways of doing this, from simply embedding html in the target web- page to interacting with the api. in this paper we will begin by looking at the simplest methods for users uncomfortable with programming, and then move on to three more advanced methods using rss, json, and xml. our examples use php, a cross-platform scripting language that may be run on either linux/ unix or windows servers. while it is not possible for us to address the many environments (such as cmses) in which websites are constructed, our code should be adaptable to most contexts. this will be especially sim- ple in the many popular php–based cmses such as drupal, joomla, and wordpress. it should be noted that the pro- cess of tagging resources in delicious requires little technical expertise, so the task of assembling lists of resources can be accomplished by any librarian. the construction of a website infrastructure (presumably by the library’s webmaster) is a more complex task that may require some programming expertise. linkrolls and tagrolls the simplest way of sharing links is to point users directly to the desired andrew darby (adarby@ithaca.edu) is web services librarian, and ron gilmour (rgilmour@ithaca.edu) is science librarian at ithaca college library, ithaca, new york. figure . delicious linkroll page adding delicious data to your library website | darby and gilmour delicious page. to share all the items labeled “biology” for the user account “iclibref,” one could disseminate the url http://delicious.com/iclibref/ biology. the obvious downside is that the user is no longer on your website, and they may be confused by their new location and what they are supposed to do there. linkrolls, a utility avail- able from the delicious site, provides a number of options for generating code to display a set of bookmarked links, including what tags to display, the number, the type of bullet, and the sorting criterion (see figure ). this utility creates simple html code that can be added to a website. a related tool, tagrolls, creates the ubiquitous delicious tag cloud. for many librarians, this will be enough. with the embedded linkroll code, and perhaps a bit of css styling, they will be satisfied with the results. however, delicious also offers more advanced methods of interacting with data. for more control over how delicious data appears on a website, the user must interact with delicious through rss, json or xml. rss like most web . applications, delicious makes its content available as rss feeds. feeds are available at a variety of levels, from the delicious system as a whole down to a par- ticular tag in a particular account. within a library context, the most useful types of feeds will be those that point to lists of resources with a given tag. for example, the request http://feeds.delicious.com/rss/icli- bref/biology returns the rss feed for the “biology” tag of the “iclibref” account, with items listed as follows: darwin’s dangerous idea (evolution ) - - t : : z http://icarus.ithaca .edu/cgi-bin/pwebrecon. cgi?bbid= iclibref this epi- sode interweaves the drama in key moments of darwin&# ;s life with documentary sequences of current research, linking past to present and introducing major concepts of evolutionary theory. biology to display delicious rss results on a website, the webmaster must use some rss parsing tool in com- bination with a script to display the results. the xml_rss package pro- vides an easy way to read rss using php. the code for such an operation might look like this: parse(); foreach ($rss->getitems() as $item) { echo “”; } ?> this code uses xml_rss to parse the rss feed and then prints out a list of linked results. rss is designed primarily as a cur- rent awareness tool. consequently, a delicious rss feed only returns the most recent thirty-one items. this makes sense from an rss perspec- tive, but it will not often meet the needs of librarians who are using delicious as a repository of resources. despite this limitation, the delicious rss feed may be useful in cases where currency is relevant, such as lists of recently acquired materials. json a second method to retrieve results from delicious is using javascript object notation or json. as with the rss feed method, a request with credentials goes out to the delicious server. the response returns in json format, which can then be processed using javascript. an example request might be http://feeds.delicious . c o m / v / j s o n / i c l i b r e f / b i o l o g y . by navigating to this url, the json response can be observed directly. a json response for a single record (formatted for readability) looks like this: delicious.posts = [ {“u”:“http:\/\/icarus.ithaca .edu\/cgi-bin\/pwebrecon .cgi?bbid= ”, “d”:“darwin’s dangerous idea (evolution )”, “t”:[“biology”], “dt”:“ - - t : : z”, “n”:“this episode interweaves the drama in key moments of darwin’s life with docu- mentary sequences of current research, linking past to present and introducing major concepts of evolutionary theory. ”} ]; it is instructive to look at the json feed because it displays the information elements that can be extracted: “u” for the url of the resource, “d” for the title, “t” for a comma-separated list of related tags, “n” for the note field, and “dt” for the timestamp. to display results in a webpage, the feed is requested using javascript: information technology and libraries | june then the json objects must be looped through and displayed as desired. alternately, as in the script below, the json objects may be placed into an array for sorting. the following is a simple exam- ple of a script that displays all of the available data with each item in its own paragraph. this script also sorts the links alphabetically. while rss returns a maximum of thirty-one entries, json allows a maximum of one hundred. the exact number of items returned may be modified through the count param- eter at the end of the url. at the ithaca college library, we chose to use json because at the time, delicious did not offer the convenient tagrolls, and the results returned by rss were displayed in reverse chronological order and truncated at thirty-one items. currently, we have a single php page that can display any delicious result set within our library website template. librarians gener- ate links with parameters that desig- nate a page title, a comma-delimited list of desired tags, and whether or not item descriptions should be displayed. for example, www.itha- calibrary.com/research/delish_feed. php?label=biology% films&tag=bio logy,biologyi¬es=yes will return a page that looks like figure . the advantage of this approach is that librarians can easily gener- ate webpages on the fly and send the url to their faculty members or add it to a subject guide or other webpage. the php script only has to read the “$_get” variables from the url and then query delicious for this content. xml delicious offers an application pro- gramming interface (api) that returns xml results from que- ries passed to delicious through https. for instance, the request https://api.del.icio.us/v /posts/ recent?&tag=biology returns an xml document listing the fifteen most recent posts tagged as “biology” for a given account. unlike either the rss or the json methods, the xml api offers a means of retrieving all of the posts for a given tag by allowing requests such as https://api.del.icio.us/v / posts/all?&tag=biology. this type of request is labor intensive for the delicious server, so it is best to cache the results of such a query for future use. this involves the user writing the results of a request to a file on the server and then checking to see if such an archived file exists before issuing another request. a php util- ity called deliciousposts, which provides caching functionality, is available for free. note that the username is not part of the request and must be sup- plied separately. unlike the public rss or json feeds, using the xml api requires users to log in to their own account. from a script, this can be accomplished using the php curl function: $ch = curl_init(); curl_setopt($ch, curlopt_ url, $queryurl); curl_setopt($ch, curlopt_ userpwd, $username . “:” . $password); curl_setopt($ch, curlopt_ returntransfer, ); $posts = curl_exec($ch); curl_close($ch); this code logs into a delicious account, passes it a query url, and makes the results of the query avail- able as a string in the variable $posts. the content of $posts can then be processed as desired to create web content. one way of doing this is to use an xslt stylesheet to transform the results into html, which can then be printed to the browser: /* create a new dom document from your stylesheet */ $xsl = new domdocument; $xsl->load(“mystylesheet.xsl”); /* set up the xslt processor */ $xp = new xsltprocessor; $xp->importstylesheet($xsl); /* create another dom docu- ment from the contents of the $posts variable */ $doc = new domdocument; $doc->loadxml($posts); /* perform the xslt transfor- mation and output the resulting html */ $html = $xp- >transformtoxml($doc); echo $html; conclusion delicious is a great tool for quickly and easily saving bookmarks. it also offers some very simple tools such as linkrolls and tagrolls to add delicious content to a website. but to exert more control over this data, the user must interact with the delicious api or feeds. we have outlined three different ways to accomplish this: rss is a familiar option and a good choice if the data is to be used in a feed reader, or if only the most recent items need be shown. json is per- haps the fastest method, but requires some basic scripting knowledge and can only display one hundred results. the xml option involves more pro- gramming but allows an unlimited number of results to be returned. all of these methods facilitate the use of delicious data within an existing website. references . delicious, tools, http://delicious .com/help/tools (accessed nov. , ). . linkrolls may be found from your delicious account by clicking settings > linkrolls, or directly by going to http:// delicious.com/help/linkrolls (accessed nov. , ). . tagrolls may be found from your delivious account by clicking settings > tagrolls or directly by going to http:// delicious.com/help/tagrolls (accessed nov. , ) . martin jansen and clay loveless, “pear::package::xml_rss,” http://pear .php.net/package/xml_rss (accessed november , ). . introducing json, http://json.org (accessed nov. , ). . ron gilmour, “deliciousposts,” h t t p : / / r o n g i l m o u r. i n f o / s o f t w a r e / deliciousposts (accessed nov. , ). lita cover , cover , cover mit press index to advertisers information technology and libraries | december information discovery insights gained from multipac, a prototype library discovery system alex a. dolski at the university of nevada las vegas libraries, as in most libraries, resources are dispersed into a number of closed “silos” with an organization-centric, rather than patron-centric, layout. patrons frequently have trouble navigating and discovering the dozens of disparate interfaces, and any attempt at a global overview of our information offerings is at the same time incomplete and highly complex. while consolidation of interfaces is widely considered to be desirable, certain challenges have made it elusive in practice. m ultipac is an experimental “discovery,” or meta- search, system developed to explore issues sur- rounding heterogeneous physical and networked resource access in an academic library environment. this article discusses some of the reasons for, and outcomes of, its development at the university of nevada las vegas (unlv). n the case for multipac fragmentation of library resources and their interfaces is a growing problem in libraries, and unlv libraries is no exception. electronic information here is scattered across our innovative webpac; our main website, our three branch library websites; remote article databases, local custom databases, local digital collections, special collections, other remotely hosted resources (such as libguides), and others. the number of these resources, as well as the total volume of content offered by the libraries, has grown over time (figure ), while access provisions have not kept pace in terms of usability. in light of this dilemma, the libraries and various units within have deployed finding and search tools that provide browsing and searching access to certain subsets of these resources, depending on criteria such as n the type of resource; n its place within the libraries’ organizational structure; n its place within some arbitrarily defined topical categorization of library resources; n the perceived quality of its content; and n its uniqueness relative to other resources. these tools tend to be organization-centric rather than patron-centric, as they are generally provisioned in relative isolation from each other without placing as much emphasis on the big picture (figure ). the result is, from the patron’s perspective, a disaggregated mass of information and scattered finding tools that, to varying degrees, each accomplishes its own specific goals at the expense of macro-level findability. currently, a compre- hensive search for a given subject across as many library resources as possible might involve visiting a half-dozen interfaces or more—each one predicated upon awareness of each individual interface, its relation to the others, and figure . “silos” in the library figure . organization-centric resource provisioning alex a. dolski (alex.dolski@unlv.edu) is web & digitization application developer at the university of nevada las vegas libraries. information discovery insights gained from multipac | dolski the characteristics of its spe- cific coverage of the corpus of library content. our library website serves as the de facto gate- way to our electronic, net- worked content offerings. yet usability studies have shown that findability, when given our website as a starting point, is poor. undoubtedly this is due, at least in part, to interface fragmentation. test sub- jects, when given a task to find something and asked to use the library website as a starting point, fail out- right in a clear majority of cases. multipac is a technical prototype that serves as an exploration of these issues. while the system itself breaks no new technical ground, it brings to the forefront critical issues of metadata quality, organizational structure, and long-term planning that can inform future actions regard- ing strategy and implemen- tation of potential solutions at unlv and elsewhere. yet it is only one of numerous ways that these issues could be addressed. in an abstract sense, multipac is biased toward principles of simplification, consolidation, and unifica- tion. in theory, usability can be improved by eliminating redundant interfaces, con- solidating search tools, and bringing together resource-specific features (e.g., opac holdings status) in one interface to the maximum extent possible (figure ). taken to an extreme, this means being able to support searching all of our resources, regardless of type or location, from a single interface; abstracting each resource from whatever native or built-in user interface it might offer; and relying instead on its data interface for querying and result-set gathering. thus multipac is as much a proof-of-concept as it is a concrete implementation. n background: how multipac became what it is multipac came about from a unique set of circumstances. from the beginning, it was intended as an exploratory project, with no serious expectation of it ever being deployed. our desire to have a working prototype ready for our discovery mini-conference meant that we had just six weeks of development time, which was hardly sufficient for anything more than the most agile of table . some popular existing library discovery systems name company/institution commercial status aquabrowser serials solutions commercial blacklight university of virginia open-source (apache) encore innovative interfaces commercial extensible catalog university of rochester open-source (mit/gpl) libraryfind oregon state university open-source (gpl) metalib ex libris commercial primo ex libris commercial summon serials solutions commercial vufind villanova university open-source (gpl) worldcat local oclc commercial table . some existing back-end search servers name company/institution commercial status endeca endeca technologies commercial idol autonomy commercial lucene apache foundation open-source (apache) search server microsoft commercial search server express microsoft free solr (superset of lucene) apache foundation open-source (apache) sphinx sphinx technologies open-source (gpl) xapian community open-source (gpl) zebra index data open-source (gpl) information technology and libraries | december development models. the resulting design, while foun- dationally solid, was limited in scope and depth because of time constraints. another option, instead of developing multipac, would have been to demonstrate an existing open-source discovery system. the advantage of this approach is that the final product would have been considerably more advanced than anything we could have developed our- selves in six weeks. on the other hand, it might not have provided a comparable learning opportunity. n survey of similar systems were its development to continue, multipac would find itself among an increasingly crowded field of competitors (table ). a number of library discovery systems already exist, most backed by open-source or commercially available back-end search engines (table ), which handle the nitty-gritty, low-level ingestion, indexing, and retrieval. these lists of systems are by no means comprehensive and do not include notable experimental or research systems, which would make them much longer. n architecture in terms of how they carry out a search, meta-search applications can be divided into two main groups: dis- tributed (or federated search), in which searches are “broadcast” to individual resources that return results in real time (figure ); and harvested search, in which searches are carried out against a local index of resource contents (figure ). both have advantages and disadvan- tages beyond the scope of this article. multipac takes the latter approach. it consists of three primary components: the search server, the user interface, and the metadata harvesting system (figure ). figure . the federated search process figure . the harvested search process figure . the three main components of multipac figure . patron-centric resource provisioning information discovery insights gained from multipac | dolski n search server after some research, solr was chosen as the search server because of its ease of use, proven library track record, and http–based representational state transfer (rest) application programming interface (api), which improves network-topological flexibility, allowing it to be deployed on a different server than the front-end web application—an important consideration in our server environment. jetty—a java web application server bundled with solr—proved adequate and convenient for our needs. the metadata schema used by solr can be customized. we derived ours from the unqualified dublin core meta- data element set (dcmes), with a few fields removed and some fields added, such as “library” and “depart- ment,” as well as fields that support various multipac features, such as thumbnail images, and primary record urls. dcmes was chosen for its combination of general- ity, simplicity, and familiarity. in practice, the solr schema is for finding purposes only, so whether it uses a standard schema is of little importance. n user interface the front-end multipac system is written in php . in a model-view-controller design based on classical object design principles. to support modularity, new resources can be added as classes that implement a resource-class interface. the multipac html user interface is composed of five views: search, browse, results, item, and list, which exist to accommodate the finding process illustrated in figure . each view uses a custom html template that can be easily styled by nonprogrammer web designers. (needless to say, judging by figures – , they haven’t been.) most dynamic code is encapsulated within dedi- cated “helper” methods in an attempt to decouple the templates from the rest of the system. output formats, like resources, are modular and decoupled from the core of the system. the html user interface is one of several interfaces available to the multipac system; others include xml and json, which effectively add web services support to all encompassed resources—a feature missing from many of the resources’ own built-in interfaces. n search view search view (figure ) is the simplest view, serving as the “front page.” it currently includes little more than a brief introduction and search field. the search field is not complicated; it is, in fact, possible to include search forms on any webpage and scope them to any subset of resources on the basis of facet queries. for example, a search form could be scoped to las vegas–related resources in special collections, which would satisfy the demand of some library departments for custom search engines tailored to their resources without contribut- ing to the “interface fragmentation” effect discussed in the introduction. (this would require a higher level of metadata quality than we currently have, which will be discussed in depth later.) because search forms can be added to any page, this view is not essential to the multipac system. to improve simplification, it could be easily removed and replaced with, for example, a search form on the library homepage. n browse view browse view (figure ) is an alternative to search view, intended for situations in which the user lacks a “concrete target” (figure ). as should be evident by its appearance, figure . the information-finding process supported by multipac figure . the multipac search view page information technology and libraries | december this is the least-developed view, simply displaying facet terms in an html unordered list. notice the facet terms in the format field; this is malprocessed, marc– encoded information resulting from a quick-and-dirty extensible stylesheet language (xsl) transformation from marcxml to solr xml. n results view the results page (figure ) is composed of three columns: . the left column displays a facet list—a feature gen- erally found to be highly useful for results-gathering purposes. the data in the list is generated by solr and transformed to an html unordered list using php. the facets are configurable; fields can be made “facetable” in the solr schema configuration file. . the center column displays results for the current search query that have been provided by solr. thumbnails are available for resources that have them; generic icons are provided for those that do not. currently, the results list displays item title and description fields. some items have very rich descriptions; others have minimal descriptions or no descriptions at all. this happens to be one of several significant metadata quality issues that will be discussed later. . the right column displays results from nonin- dexed resources, including any that it would not be feasible to index locally, such as google, our article databases, and so on. multipac displays these resources as collapsed panes that expand when their titles are clicked and initiate an ajax request for the current search query. in a situation in which there might be twenty or more “panes” to load, performance would obviously suffer greatly if each one had to be queried each time the results page loaded. the on-demand loading process greatly speeds up the page load time. currently, the right column includes only a handful of resource panes—as many as could be developed in six weeks alongside the rest of the prototype. it is anticipated that further development would entail the addition of any number of panes—perhaps several dozen. the ease of developing a resource pane can vary greatly depending on the resource. for developer- friendly resources that offer a useful javascript object notation (json) api, it can take less than half an hour. for article databases, which vendors generally take great pains to “lock down,” the task can entail a two-day marathon involving trial-and-error http-request-token authentication and screen-scraping of complex invalid html. in some cases, vendor license agreements may prohibit this kind of use altogether. there is little we can do about this; clearly, one of multipac’s severest limita- tions is its lack of adeptness at searching these types of “closed” remote resources. n item view item view (figure ) provides greater detail about an individual item, including a display of more metadata fields, an image, and a link to the item in its primary con- text, if available. it is expected that this view also would include holdings status information for opac resources, although this has not been implemented yet. the availability of various page features is dependent on values encoded in the item’s solr metadata record. for example, if an image url is available, it will be displayed; if not, it won’t. an effort was made to keep the view logic separate from the underlying resource to improve code and resource maintainability. the page template itself does not contain any resource-dependent conditionals. n list view list view (figure ), essentially a “favorites” or “cart” view, is so named because it is intended to duplicate the list feature of unlv libraries’ innovative millennium figure . the multipac browse view page information discovery insights gained from multipac | dolski opac. the user can click a button in either results view or item view to add items to the list, which is stored in a cookie. although currently not feature-rich, it would be reasonable to expect the ability to send the list as an e-mail or text message, as well as other features. n metadata harvesting system for metadata to be imported into solr, it must first be harvested. in the harvesting process, a custom script checks source data and com- pares it with local data. it downloads new records, updates stale records, and deletes missing records. not all resources support the ability to easily check for changed records, meaning that the full record set must be down- loaded and converted during every harvest. in most cases, this is not a problem; most of our resources (the library catalog excluded) can be fully dumped in a matter of a few seconds each. in a production environment, the harvest scripts would be run automatically every day or so. in practice, every resource is different, necessitating a different harvest script. the open archives initiative protocol for metadata harvesting (oai-pmh) is the proto- col that first jumps to mind as being ideal for metadata harvesting, but most of our resources do not support it. ideally, we would modify as many of them as possible to be oai–compliant, but that would still leave many that are out of our hands. either way, a substantial number of custom harvest scripts would still be required. for demonstration purposes, the multipac prototype was seeded with sample data from a handful of diverse resources: . a set of , marc records from our library catalog, which we converted to marcxml and then to solr xml using xsl transformations . our locally built las vegas architects and buildings database, a mysql database containing more than , rows across tables, which we queried and dumped into xml using a php script . our locally built special collections database, a smaller mysql database, which we dealt with the same way . our contentdm digital collections, which we downloaded via oai-pmh and transformed using another custom xsl stylesheet there are typically a variety of conversion options for each resource. because of time constraints, we simply chose what we expected would be the quickest route for each, and did not pay much attention to the quality of the conversion. n how multipac answers unlv libraries’ discovery questions multipac has essentially proven its capability of solv- ing interface multiplication and fragmentation issues. figure . the multipac results view page information technology and libraries | december by adding a layer of abstraction between resource and patron, it enables us to reference abstract resources instead of their specific implementations—for example, “the library catalog” instead of “the innopac catalog.” this creates flexibility gains with regard to resource pro- vision and deployment. this kind of “pervasive decoupling” can carry with it a number of advantages. first, it can allow us to provide custom-developed services that vendors cannot or do not offer. second, it can prevent service interruptions caused by maintenance, upgrades, or replacement of individual back-end resources. third, by making us less dependent on specific implementations of vendor products—in other words, reducing vendor “lock-in”—it can potentially give us leverage in vendor contract negotiations. because of the breadth of information we offer from our website gateway, we as a library are particularly sensitive about the continued availability of access to our resources at stable urls. when resources are not persistent, patrons and staff need to be retrained, expec- tations need to be adjusted, and hyperlinks—scattered all over the place—need to be updated. by decoupling abstract resources from their implementations, multipac becomes, in effect, its own persistent uri system, unify- ing many library resources under one stable uri schema. in conjunction with a url rewriting system on the web server, a resource-based uri schema (figure ) would be both powerful and desirable. n lessons learned in the development of multipac the lessons learned in the development of multipac fall into three main categories, listed here in order of importance. metadata quality considerations quality metadata—characterized by unified schemas; useful crosswalking; and consistent, thorough descrip- tion—facilitates finding and gathering. in practice, a sur- rogate record is as important as the resource it describes. below a certain quality threshold, its accompanying resource may never be found, in which case it may as well not exist. surrogate record quality influences relevance ranking and can mean the difference between the most relevant result appearing on page or page (relevance, of course, being a somewhat disputed term). solr and similar systems will search all surrogates, including those that are of poor quality, but the resulting relevancy rank- ing will be that much less meaningful. figure . example of an implementation-based vs. resource-based uri implementation-based http://www.library.unlv.edu/arch/archdb /index.php/projects/view/ resource-based (hypothetical) http://www.library.unlv.edu/item/ figure . the multipac item view page figure . the multipac list view page information discovery insights gained from multipac | dolski metadata quality can be evaluated on several lev- els, from extremely specific to extremely broad (figure ). that which may appear to be adequate at one level may fail at a higher level. using this figure as an example, multipac requires strong adherence to level , whereas most of our metadata fails to reach level . a “level failure” is illustrated in table , which compares sample metadata records from four different multipac resources. empty cells are not necessarily “bad”— not all metadata elements apply to all resources—but this type of inconsistency multiplies as the number of resources grows, which can have negative implications for retrieval. suggestions for improving metadata quality the results from the multipac project suggest that meta- data rules should be applied strictly and comprehensively according to library-wide standards that, at our libraries, have yet to be enacted. surrogate records must be treated as must-have (rather than nice-to-have) features of all resources. resources that are not yet described in a system that supports search- able surrogate records should be transitioned to one that does; for example, html web- pages should be tran- sitioned to a content management system with metadata ascrip- tion and searchability features (at unlv, this is planned). however, it is not enough for resources to have high-quality meta- data if not all schemas are in sync. there exist a number of resources in our library that are well-described but whose schemas do not mesh well with other resources. different formats are used; different descriptive elements figure . example scopes of metadata application and evalua- tion, from broad (top) to specific table . comparing sample crosswalked metadata from four different unlv libraries resources library catalog digital collections special collections database las vegas architects & buildings database title goldfield: boom town of nevada map of tonopah mining district, nye county, nevada : mines and mining collection flamingo hilton las vegas creator paher, stanley w. booker & bradford call number f .g p contents (item-level description of contents) format digital object photo collections database record language eng eng eng coverage tonopah mining district (nev.) ; ray mining district (nev.) description (omitted for brevity) publisher nevada publications university of nevada las vegas libraries unlv architecture studies library subject (lcsh omitted for brevity) (lcsh omitted for brevity) information technology and libraries | december are used; and different interpretations, however subtle, are made of element meanings. despite the best intentions of everyone involved with its creation and maintenance, and despite the high quality of many of our metadata records when examined in isola- tion, in the big picture, multipac has demonstrated—per- haps for the first time—how much work will be needed to upgrade our metadata for a discovery system. would the benefits make the effort worthwhile? would the effort be implementable and sustainable given the limitations of the present generation of “silo” systems? what kind of adjustments would need to be made to accommodate effective workflows, and what might those workflows look like? these questions still await answers. of note, all other open-source and vendor systems suffer from the same issues, which is a key reason that these types of systems are not yet ascendant in libraries. there is much promise in the ability of infrastructural standards like frbr, skos, rda, and the many other esoteric information acronyms to pave the way for the next generation of library discovery systems. organizational considerations electronic information has so far proved relatively elusive to manage; some of it is ephemeral in existence, most of it is constantly changing, and all of it is from diverse sources. attempts to deal with electronic resources—representing them using catalog surrogate records, streamlining web- site portals, farming out the problem to vendors—have not been as successful as they have needed to be and suf- fer from a number of inherent limitations. multipac would constitute a major change in library resource provision. our library, like many, is for the most part organized around a core s– s ils–support model that is not well adapted to a modern unified discovery environment. next-generation discovery is trending away from assembly-line-style acquisition and processing of primarily physical resources and toward agglomerating interspersed networked and physical resource clouds from on- and offsite. in this model, increasing responsibilities are placed on all content pro- viders to ensure that their metadata conforms to site-wide protocols that, at our library, have yet to be developed. n conclusion in deciding how to best deal with discovery issues, we found that a traditional product matrix comparison does not address the entire scope of the problem, which is that some of the discoverability inadequacies in our libraries are caused by factors that cannot be purchased. sound metadata is essential for proper functioning of a unified discovery system, and descriptive uniformity must be ensured on multiple levels, from the element level to the institution level. technical facilitators of improved discoverability already exist; the responsibility falls on us to adapt to the demands of future discovery systems. the specific discovery tool itself is only a facilitator, the specific implementation of which is likely to change over time. what will not change are library-wide metadata quality issues that will serve any tool we happen to deploy. the multipac project brought to light important library-wide discoverability issues that may not have been as obvious before, exposing a number of limitations in our exist- ing metadata as well as giving us a glimpse of what it might take to improve our metadata to accommodate a next-generation discovery system, in whatever form that might take. references . unlv libraries usability committee, internal library website usability testing, las vegas, . . karen calhoun, “the changing nature of the catalog and its integration with other discovery tools.” report prepared for the library of congress, . . xiaoming liu et al., “federated searching interface tech- niques for heterogeneous oai repositories,” journal of digital information , no. ( ). . apache software foundation, apache solr, http://lucene .apache.org/solr/ (accessed june , ). . dublin core metadata initiative, “dublin core metadata element set, version . ,” jan. , , http://dublincore.org/ documents/dces/ (accessed june , ). . lorcan dempsey, “a palindromic ils service layer,” lorcan dempsey’s weblog, jan. , , http://orweblog.oclc .org/archives/ .html (accessed july , ). . tod a. olson, “utility of a faceted catalog for scholarly research,” library hi tech , no. ( ): – . . tim berners-lee, “hypertext style: cool uris don’t change,” , http://www.w .org/provider/style/uri (accessed june , ). . bowen, jennifer, “metadata to support next-generation library resource discovery: lessons from the extensible cata- log, phase ,” information technology and libraries , no. (june ): – . . calhoun, “the changing nature of the catalog.” information technology and libraries | september andrew k. pacepresident’s message andrew k. pace (pacea@oclc.org) is lita president / and executive director, networked library services at oclc inc. in dublin, ohio. w elcome to my first ital column as lita presi- dent. i’ve had the good fortune to write a number of columns in the past—in computers in libraries, smart libraries newsletter, and most recently american libraries—and it is a role that i have always cherished. there is just enough space to say what you want, but not all the responsibility of backing it up with facts and figures. in the past, i have worried about hav- ing enough to say month after month for an undefined period. now i am daunted by only having one year to address the lita membership and communicate goals and accomplishments of my quickly passing tenure. i am simultaneously humbled and extremely excited to start my presidential year with lita. i have some ambitious agenda items for the division. i said when i was running that i wanted to make lita the kind of orga- nization that new librarians and it professionals want to join and that seasoned librarians wanted to be active in. recruitment to lita is vital, but there is also work to be done to make that recruitment even easier. i am fortunate in following up the great work of my predecessors, many of whom i have had the pleasure of serving with on the lita board since . they have set the bar for me and make the coming year as challenging as anything i have done in my career. i also owe a lot to the membership who stepped forward to volunteer for committees, liaison appointments, and other volunteer opportunities. i also think it is important for lita mem- bers to know just how much the board relies on the faith- ful and diligent services of the lita staff. at my vice presidential town meeting, i talked about marketing and communication in terms of list (who), method (how), and message (what and why). not only was this a good way to do some navel gazing on what it means to be a member of lita, it laid some groundwork for the year ahead. i think it is an inescapable conclusion that the lita board needs to take another look at stra- tegic planning (which expires this year). the approach i am going to recommend, however, is not one that tries to connote the collective wisdom of a dozen lita leaders. instead, i hope we can define a methodology by which lita committees, interest groups, and the membership at large are empowered to both do the work of the division and benefit from it. one of the quirky things that some people know about me is that i actually love bureaucracy. i was pleased to read in the lita bylaws that it is actually my duty as president to “see that the bylaws are observed by the officers and members of the board of directors.” i will tell you all that i also interpret this to mean that the president and the board will not act in ways that are not prescribed. the strength of a volunteer organization comes from its volunteers. the best legacy a lita president can provide is to give committees, interest groups, and the member- ship a free reign to create its future. as for the board, its main objective is to oversee the affairs of the division during the period between meet- ings. frankly, we’re not so great at this, and it is one of the biggest challenges for any volunteer organization. it is also one of my predecessor’s initiatives that i plan to follow through on with his help as immediate past presi- dent. participation and involvement—and the ability to follow the work and strategies of the division—should be easier for all of us. so, if i were to put my platform in a nutshell it would be this—recruitment, communication, strategic planning, and volunteer empowerment. i left out fun, because it goes without saying that most of us are part of lita because it’s a fun division with great members. this is a lot to get done in one year, but because it will be fun, i’m looking forward to it. information technology and libraries | june an assessment of student satisfaction with a circulating laptop service louise feldmann, lindsey wess, and tom moothart since may of , colorado state university’s (csu) morgan library has provided a laptop computer lending service. in five years the service had expanded from to laptops. although the service was deemed a success, users complained about slow laptop startups, lost data, and lost wireless connections. in the fall of , the pro- gram was formally assessed using a customer satisfaction survey. this paper discusses the results of the survey and changes made to the service based on user feedback. colorado state university (csu) is a land-grant insti-tution located in fort collins, colorado. the csu libraries consist of the morgan library, the main library on the central campus; the veterinary teaching branch hospital library at the veterinary hospital cam- pus; and the atmospheric branch library at the foothills campus. in , morgan library completed a major reno- vation and expansion which provided a designated space for public desktop computers in an information commons environment. the library called this space the electronic information center (eic). due to the popularity of the eic ,and with the intent of expanding computer access without expanding the computer lab, library staff began to explore the implementation of a laptop checkout service in . library staff used heather lyle’s ( ) arti- cle “circulating laptop computers at west virginia university” as a guide in planning the service. development funds were used to purchase twenty laptop computers, and the com corporation donated fifteen wireless network access points. the laptops were to be used in morgan library on a wireless network main- tained by the library technology services department. these computers were to be circulated from the loan desk, the same desk used to check out books. although the building is open to the public, use of the laptops was limited to university students and staff and for library in-house use only. all the public desktop computers and laptops use microsoft windows and microsoft office. maintaining the security of the libraries’ network and students’ personal data in a wireless environment was paramount. to maintain a secure computing environment and present a standardized computing experience in the library, an application of windows xp group policies was used. currently, the laptop software is updated at least every semester using symantec ghost. ghost cop- ies a standardized image to every laptop even when the library owns a variety of computer models from the same manufacturer. additionally, due to concerns over wireless computer security, morgan library implemented cisco’s virtual private network (vpn) in . the laptop service was launched in may . more than , laptop transactions occurred in the initial year. since its inception, the use of the morgan library laptop service and the number of laptops available for checkout has steadily grown. using student technology funds, the service had grown to laptops and ten presenta- tion kits consisting of a laptop, projector, and a portable screen. circulation during the fall semester totaled , laptops and presentation kits. in fiscal year , , laptops and presentation kits were checked out. based on the high circulation statistics and anecdotal evidence, the service appeared to be successful. although morgan library replaced laptops every three years and upgraded the wireless network, laptop support staff noted that users complained of slow laptop startups, lost data, and lost wireless connections. the researchers also noted that large numbers of users queued at the circulation desk at : p.m. even though large numbers of desktop computers were available in the eic. a customer service satisfaction survey was developed to assess the service and test library staff’s assumptions about the service. csu had a student population of , students at the time of the survey. n literature review much of the published literature discussing laptop ser- vices focuses on the implementation of laptop lending pro- grams and was published from to , when many libraries were beginning this service (allmang ; block ; dugan ; myers ; oddy ; vaughan and burnes ; williams ). these articles deal primarily with topics such as how to deal with start-up technologi- cal, staffing, and maintenance issues. they have minimal discussion of the service post-implementation. researchers who have surveyed users of university laptop lending services include direnzo ( ), lyle ( ), jordy ( ), block ( ), oddy ( ), and monash university’s caulfield library ( ). direnzo from the university of akron only briefly discusses a survey they conducted with some information about additional software added as a result of their user com- ments. lyle from west virginia university discusses the percentage of respondents to particular questions such louise feldmann (louise.feldmann@colostate.edu) is the business and economics librarian at colorado state university libraries. she serves as the college liaison librarian to the college of business. lindsey wess (lindsey.wess@colostate. edu) coordinates assistive technology services and manages the information desk and the electronic information center at colorado state university libraries. tom moothart (tmoothar@ library.colostate.edu) is the coordinator of on-site services at colorado state university libraries. student satisfaction with circulating laptop service | feldmann, wess, and moothart as what applications were used, problems encountered, and overall satisfaction with the service. jordy’s report provides in-depth analysis of the survey results from the university of north carolina at chapel hill, but the focus of his survey is on the laptop service’s impact on library employee work flow. monash university’s caulfield library survey focuses on wireless access and awareness of the program by patrons. other survey results found on university library web sites include southern new hampshire university library (west ) and murray state university library ( ). additionally, the monmouth university library web site ( ) provides discussion and written analysis of a survey they conducted prior to implementation of their service, a survey which was used to gather informa- tion and assess patron needs in order to aid in the con- struction and planning of their service. from the survey results discussed in the literature and posted on web sites, overall comments from users are very consistent with one another. most users indicate that they use a loaned laptop computer rather than desktop computer for privacy and portability (lyle ; oddy ; west ). in addition, the responses from patrons are overwhelmingly positive and users appreciated hav- ing the service made available (lyle ; jordy ; west ). both west virginia university and the university of north carolina at chapel hill surveys found that percent of respondents would check out a laptop again (lyle ; jordy ). southern university of new hampshire’s survey indicated that percent of those responding would check one out again (west ). many respondents stated that a primary drawback of using the laptops was the slowness of connectivity (lyle ; monash ; murray state ). the primary use of the laptops, reported in the surveys, was microsoft word (lyle ; jordy ; oddy ). there is a lack of published literature regarding laptop lending customer satisfaction surveys and analy- sis. this could be due to the relative newness of many programs, the lack of university libraries that provide laptops, or the reliance on circulation statistics solely to assess the program. articles that discuss circulation and usage statistics as an assessment indicator to judge the popularity of their programs include direnzo ( ), dugan ( ), and vaughan and burnes ( ). based on high circulation statistics and positive anecdotal evi- dence, it may appear that library users are pleased with laptop programs, and perhaps there has been a hesita- tion to survey users on a program that is perceived by those in the library as successful. n results with the strong emphasis on assessment at colorado state university, it was decided to formally survey laptop users on their satisfaction with the program. the survey was distributed by the access services staff when the laptops were checked out from october , , to november , . this was a voluntary survey and the respondents were asked to complete one survey. users returned completed surveys. undergraduates are the predominant audience for the laptop service; of the returned sur- veys, identified themselves as undergraduates. as shown in table , the responses indicated that the library has a core of regular laptop users, with percent using the laptops at least daily and percent using the laptops at least weekly. only percent indicated that they were using a laptop for the first time. many laptop users also utilized the eic with percent responding that they use the information commons at least weekly (see table ). the laptops were initially purchased with the intent that they would be used to support student team projects. presentation kits with a laptop, projector, and portable screen were an extension of this idea and were also made available for checkout. surprisingly, only percent of table . how often do you use a library laptop? frequency percentage more than once a day % daily % weekly % monthly % my first time % n= table . how often do you use a library pc? frequency percentage more than once a day % daily % weekly % monthly % never % n= information technology and libraries | june the respondents noted that they were using the laptop with a group. during evenings, it was observed by staff that stu- dents were regularly queuing and waiting for a laptop even though pcs were available in the library computer lab. figure shows an hourly use statistics for the desk- top and laptop public computers. the usage of the desk- top computer drops in the late afternoon, just as the use of the laptop computer increases. students were asked why they chose a laptop rather than a library pc and were allowed to choose from multiple answers. as can be seen in table , most students noted the advantages of portability and privacy. five respon- dents wrote in the “other” category that they were able to work better in quieter areas, and ten mention that the computer lab workspace is limited. the dense use of space in the library computer lab has been noted by morgan library staff and students. the desktop surround- ing each library pc only provides about three feet of workspace. one respondent explained the choice of laptop over pc was because “i can take it to a table and spread out my notes vs. on a library pc.” for many users, the desktops are too crowded to spread research mate- rial, and the eic is too noisy for contemplative thought. as can be noted from the use statistics, the public laptop program has been a very popu- lar library service. prior to the survey, the per- ception of the morgan library staff was that students were waiting in the evening for extended periods of time for a lap- top. when the library expanded the laptop pool from in to in , it had seemingly no effect on reducing the number of students wait- ing to use them. as can be seen in table , when asked how long they had waited for a laptop, percent of the students said they had access to a laptop immediately, and percent waited less than a minute. the survey was administered during the second busiest time of the year for the library, the month before thanksgiving break. in the open comments, one respondent stated that it was possible to wait forty- five minutes to an hour for a laptop and another noted that “during finals weeks it is almost impossible to get one.” even with the limited waiting time recorded by the page of feldmann figures.doc % % % % % % % % % % % : am : am : am : am : am : pm : pm : pm : pm : pm : pm : pm : pm : pm : pm : pm : pm time of day p er ce nt ag e of u se r desktop computers checkout laptops figure . computer use statistics for may , . figure . computer use statistics for may , . table . why did you choose to use a laptop rather than a library pc? response number portability privacy easier to work with a group portability and privacy portability and easier to work with a group portability, privacy, and easier to work with a group student satisfaction with circulating laptop service | feldmann, wess, and moothart respondents, when asked how the library could improve the laptop service many respondents requested that more laptops be purchased to decrease the wait. the library is struggling to determine the appropriate number of lap- tops to have available during peak use periods to reduce or eliminate wait times. the library laptops are more problematic than the library desktop computers to support. the laptops are more fragile than the desktop computers and have the added complication of connecting to the wireless net- work. every morning the morgan library’s technology staff retrieves non-functioning laptops; library technicians regularly retrieve lost data due to malfunctioning laptops and unsophisticated computer users. the addition of the virtual private network (vpn) connection to the laptop startup script files has slowed the boot-up to the wireless network. an effort has been made to ameliorate wireless “dead zones,” but users still complain of being dropped from the wireless network. with these problems in mind, users were asked about the technical complications they have experienced with the library laptops. the survey responses in tables and indicate a much lower percentage of users reporting technical problems than was anticipated. the technical staff’s large volume of technical calls may reflect the volume of users rather than systematic problems with the laptop service. surprisingly, percent of the users reported rarely or never returning a non-functioning laptop. in addition, the library technicians have reported that no problems have been found on some of the laptops returned for repair. some of the returned computers may be due to frustra- tion with the slow connection to the wireless network. forty-five percent of respondents reported at least occasionally having problems connecting to the wireless network. from the inception of the laptop program, the library has experienced problems with the wireless tech- nology. from its original fifteen wireless access points to its current twenty-nine, the library has struggled to meet the demand of additional library laptops and users’ personal laptops. many written comments on the surveys complained about the slow connection speed of the wire- less network such as, “find a way to make the boot-up process faster. i need to wait about five minutes for it to be totally booted and ready to use.” even with the slow connection to the wireless net- work, percent of students responding to the survey rated their satisfaction with the library’s laptop service as excellent and percent rated their satisfaction as good (see table ). n discussion even with percent of our users rating the laptop service as good or excellent, the survey noted some problems that needed attention. the morgan library laptops seamlessly connect to a wireless network through a login script when the computer is turned on. a new script was written to table . how long did you wait before you were able to check out your laptop? response percentage i did not wait % less than one minute % one to four minutes % five to ten minutes % more than ten minutes % n= table . how often have you experience problems saving files, connecting to the wireless network, or had a laptop that locked up or crashed? frequency saving files wireless connection locked up or crashed often < % % < % occasionally % % % rarely % % % never % % % n= table . how often have you returned a library laptop that was not working properly? frequency percentage often % occasionally % rarely % never % n= information technology and libraries | june allow the connection and authentication to the cisco virtual private network (vpn) client. during testing it was found that some laptops took as long as ten min- utes to connect to the wireless network, which resulted in numerous survey respondents commenting on our slow wireless network. to help correct this problem, the library’s network staff changed each laptop’s user profile from a mandatory roaming profile to a local profile and simplified the login script. the laptops connected faster to the wireless network with the new script, but they still did not meet the students’ expectations. in the fall of , the library network staff moved the laptops from vpn to wi-fi protected access (wpa) wireless security, and laptop login time to the wireless network dropped to under two minutes. the number of customer complaints dropped dramatically after implementing wpa. additional access points were purchased to improve connectivity in morgan library’s wireless “dead zones.” in january , the university’s central computing services audited the wireless network after continued wireless connectivity complaints. the audit recom- mended reconfiguring the access points channel assign- ments. in many cases it was found that the same channel had been assigned to access points adjacent to each other, ultimately compromising laptop connectivity. the audit also discovered noise interference on the wireless net- work from a . -ghz cordless phone used by the loan desk staff. the phone was replaced with a . -ghz one, which has resulted in fewer dropped connections near the loan desk. supporting almost laptops has introduced several problems in the library. the morgan library building was not designed to support the use of large numbers of laptops. because it is impractical for the loan desk to charge nearly laptop batteries throughout the day, laptops available for checkout must be connected to electrical outlets. these are seldom near study tables, and students are forced to crawl underneath tables to locate power or stretch adapter cords across aisles. a space plan for the morgan library is being developed that will increase the number of outlets near study tables. in the meantime, power strips were added to tables used heavily by laptop users. the loan desk staff is very efficient at circulating, but has less success at troubleshooting technical problems. when the laptop service was first implemented, large numbers of laptops were not available due to servicing reasons. the public laptop downtime was lowered by hiring additional library technology students. a one-day onsite repair service agreement was purchased from the manufacturer which resulted in many equipment repairs being completed within hours. in order to reduce the downtime further, a plan to replace some loan desk student workers with library technology students is being evaluated. the technology students will be able to troubleshoot connectivity and hardware problems with the users when they return the defective computers to the loan desk. if a computer needs additional service, it can be handled immediately, which will allow more laptops for checkout since fewer will be removed for repair. when the laptop service was first envisioned, it was seen as a great service for those working in groups. as can be seen in table , very few students are using the laptops in a group setting. in survey written comments, students emphasize that they enjoy the portability and privacy enabled by using a laptop. the morgan library eic is cramped and noisy, with the configuration allow- ing very little room for students to spread out research materials and notes for writing. the morgan library space plan takes these issues into consideration and rec- ommends reconfiguring the eic to lessen the noise and provide writing space near computers. this is intended to improve the student library experience and encourage students to use the desktop computers during the eve- nings when lines form for the laptops. in order to decrease the current laptop queue at the loan desk, more laptops will be added. as a result of survey comments requesting apple computers, five mac powerbooks were added to the library’s laptop fleet. in addition, as morgan library adds more checkout laptops and the number of students arriving on campus with wireless laptops increases, the wireless infrastructure will need to be upgraded. upgrading the wireless access points to standard . g has been implemented. updating each laptop with a new hardrive image has become problematic as the number of laptops has increased. the wireless network capacity is not large enough for the ghost software to transmit the image to multiple laptops, and so each laptop must be physically attached to the library network. initially, when library technology services attempted imaging many laptops at once, it took six to eight hours and required up to eight staff members. this method of large-scale laptop imaging was so network intensive that it had to be per- formed when the library was closed to avoid disrupting table . please rate your satisfaction with the laptop service. response percentage excellent % good % neutral % poor very poor % < % n= student satisfaction with circulating laptop service | feldmann, wess, and moothart public internet use. now imaging the laptop fleet is done piecemeal, twenty to thirty laptops at a time, in order to minimize complications with the ghost process and mul- ticasting through the network switches. due to the staff time required, laptop software is not updated as often as the users would like. technological solutions continue to be investigated that will decrease the labor and network intensity of imaging. n conclusion the morgan library laptop service was established in and has been a very popular addition to the library’s services. as an example of its popularity, in fiscal year the laptops circulated , times. student govern- ment continues to support the use of student technology fees to support and expand the fleet of laptops. this survey was an attempt to assess users’ perceptions of the service and identify areas that need improvement. the survey found that students rarely wait more than a few minutes for a laptop, and in open-ended survey ques- tions, students noted that they waited for computers only during peak use periods. while relatively few survey respondents experienced technical difficulties with the laptops and wireless network, slow wireless connection time was a concern that students noted in the open com- ments section of the survey. overall, the students gave the laptop service a very high rating. when asked to suggest improvements to the service, many respondents recommended purchasing more laptops. the libraries made several changes to improve the laptop service based on survey responses. changes have been made to the login script files, wireless network, and security protocol to speed and stabilize the wireless con- nection process. additional wireless access points will be added to the building and all access points will be upgraded to the . g standard. in addition, five mac powerbooks have been added to the fleet of windows- based laptops. the library continues to investigate new service models to circulate and maintain the laptops. works cited allmang, nancy. . our plan for a wireless loan service. com- puter in libraries , no. : – . block, karla j. . laptops for loan: the experience of a multi- library project. journal of interlibrary loan, document delivery, and information , no. : – . direnzo, susan. . a wireless laptop-lending program: the university of akron experience. technical services quarterly , no. : – . dugan, robert e. . managing laptops and the wireless net- work at the mildred f. sawyer library. journal of academic librarianship , no. : – . jordy, matthew l. . the impact of user support needs on a large academic workflow as a result of a laptop check-out program. master’s thesis, university of north carolina. lyle, heather. . circulating laptop computers at west vir- ginia university. information outlook , no. : – . myers, penelope. . laptop rental program, temple univer- sity libraries. journal of interlibrary loan, document delivery, and information supply , no. : – . monash university caulfield library. . laptop users and wireless network survey. www.its.monash.edu.au/staff/net- works/wireless/review/caul-lapandnetsurvey.pdf (accessed june , ). monmouth university. . testing the wireless waters: a sur- vey of potential users before the implementation of a wireless notebook computer lending program in an academic library. http://bluehawk.monmouth.edu/~hholden/wwl/wire- less_survey_results.html (accessed june , ). murray state university. . library laptop computer usage survey results. www.murraystate.edu/msml/laptopsurv. htm (accessed june , ). oddy, elizabeth carley. . laptops for loan. library and infor- mation update , no. : – . vaughn, james b., and brett burnes. . bringing them in and checking them out: laptop use in the modern academic library. information technology and libraries , no. : – . west, carol. . librarians pleased with results of student survey. southern new hampshire university. www.snhu. edu/ /asp (accessed june , ). williams, joe. . taming the wireless frontier: pdas, tablets, and laptops at home on the range. computers in libraries , no. : – , – . information technology and libraries | september mireia ribera turróeditorial board thoughts the june issue of ital featured a new column enti-tled editorial board thoughts. the column features commentary written by ital editorial board mem- bers on the intersection of technology and libraries. in the june issue kyle felker made a strong case for gerald zaltman’s book how customers think as a guide to doing user-centered design and assessment in the context of limited resources and uncertain user needs. in this col- umn i introduce another factor in the library–it equation, that of rapid technological change. in the midst of some recent spring cleaning in my library i had the pleasure of finding a report documenting the current and future it needs of purdue university’s hicks undergraduate library. the report is dated winter . the following summarizes the hicks undergraduate library’s it resources in : [the library] has seven public workstations running eight different databases and using six different search software programs. six of the stations support a single database only; one station supports one cd-rom application and three other applications (installed on the hard drive). none of the computers runs windows, but the current programs do not require it. five sta- tions are equipped with six-disc cd-rom drives. we do not anticipate that we will be required to upgrade to windows capability in the near future for any of the application programs. today the hicks undergraduate library’s it resources are dramatically different. as opposed to seven pub- lic workstations, we have more than seventy comput- ers distributed throughout the library and the digital learning collaboratory, our information commons. this excludes forty-six laptops available for patron checkout and eighty-eight laptops designated for instructional use. we have moved from eight cd-rom databases to more than four hundred networked databases accessible throughout the purdue university libraries, campus, and beyond. as a result, there are hundreds of “search software programs”—doesn’t that phrase sound odd today?—including the library databases, the catalog, and any number of commercial search engines like google. today all, or nearly all, of our machines run windows, and the macs have the capability of running windows. in addition to providing access to databases, our machines are loaded with productivity and multimedia software allowing students to consume and produce a wide array of information resources. beyond computers, our library now loans out additional equipment including hard drives, digital cameras, and video cameras. the report also includes system specifications for the computers. these sound quaint today. of the seven computers six were machines with processors clock- ing in at mhz. the computers had between k and . mb of ram with hard drives with capacities between and mb. the seventh computer was a machine probably with a . mhz processor, and correspond- ingly smaller memory and hard disc capacity. the report does not include monitor specifications, though, based on the time, they were likely fourteen- or fifteen-inch cga or ega cathode ray tube monitors. modern computers are astonishingly powerful in comparison. according to a member of our it unit, the computers we order today have . ghz dual core processors, gb of ram, and gb hard drives. this equates to being times faster, , times more ram, and hard drives that are , times larger than the computers! as a benchmark, consider moore’s law, a doubling of capacitors every two years, a sixty-four fold increase over a thirteen year period. who would have thought that library computers would outpace moore’s law?! today’s computers are also smaller than those of . our standard desktop machines serve as an example, but perhaps not as dra- matically as laptops, mini-laptops, and any of the mobile computing machines small enough to fit into your pocket. monitors are smaller, though also bigger. each new com- puter we order today comes standard with a twenty-inch flat panel lcd monitor. it is smaller in terms of weight and overall size, but the viewing area is significantly larger. these trends are certainly not unique to purdue. nearly every other academic library could boast similar it advancements. with this in mind, and if moore’s law continues as projected, imagine the computer resources that will be available on the average desktop machine— although one wonders if it will in fact be a desktop machine—in the next thirteen years. what things out on the distant horizon will eventually become com- monplace? here the quote from the report about windows is particularly revealing. what things that are currently state-of-the-art will we leave behind in the next decade? what’s dos? what’s a cd-rom? will we soon say, what’s a hard drive? what’s software? what’s a desktop computer? in the last thirteen years we have also witnessed the widespread adoption and proliferation of the internet, the network that is the backbone for many technolo- gies that have become essential components of physical and digital libraries. earlier this year, i co-authored an arl spec kit entitled social software in libraries. the survey reports on the usage of ten types of social soft- ware within arl libraries: ( ) social networking sites like myspace and facebook; ( ) media sharing sites like information technology and libraries | september matthew m. bejune (mbejune@purdue.edu) is an ital editorial board member ( – ), assistant professor of library science at purdue university, and doctoral student in the graduate school of library and information science at the university of illinois at urbana–champaign. matthew m. bejune editorial board thoughts | bejune youtube and flickr; ( ) social book- marking and tagging sites like del. icio.us and librarything; ( ) wikis like wikipedia and library success: a best practices wiki; ( ) blogs; ( ) rss used to syndicate content from webpages, blogs, podcasts, etc.; ( ) chat and instant messenger services; ( ) voice over internet protocol (voip) services like googletalk and skype; ( ) virtual worlds like second life and massively multiplayer online games (mmogs) like world of warcraft; and ( ) wid- gets either developed by libraries like facebook applications, firefox catalog search extensions, or wid- gets implemented by libraries like meebome and firefox plugins. of the arl libraries that responded, a % response rate, ( % of respondents) said they are using social software. of the three librar- ies not using social software, two indicated they plan to do so in the future. in combination then, out of respondents ( %) indicated they are either currently using or planning to use social software. as part of the survey there was a call for examples of social software used in libraries. of the examples we received, we selected around for publication in the spec kit. the examples are captivating and they illus- trate the wide variety of applications in use today. of the ten social software applications in the spec kit, how many of them were at our disposal in ? by my count three: chat and instant messenger services, voip, and virtual worlds such as text-based muds and moos. of these three, how many were in use in librar- ies? very few, if any. in our survey we asked libraries for the year in which they first implemented social software. the earliest applications were cu-seeme, a voip chat service at cornell university in , im at the university of california riverside in as well, and interoffice chat at the university of kentucky in . the remain- ing libraries adopted social software in year and beyond, with being the most common year with responses or % of the libraries that had adopted social software. a look at this data shows that my earlier use of a thirteen-year time period to illustrate how difficult it is to project technological innovations that may prove disruptive to our organizations is too broad a time frame. perhaps we should scale this back to looking at five-year increments of time. using the spec kit data, in year , a total of arl libraries had adopted social software. this represents % of the total number of institutions that responded when we did our survey. this seems like figure . responses to the question, “please enter the year in which your library first began using social software” (n= ). a more reasonable time frame to be looking to the future. so, what does the future hold for it and libraries, whether it be thirteen or five years in the future? i am not a technologist by training, nor do i consider myself a futurist, so i typically defer to my colleagues. there are three places i look to for prognostications of the future. the first is lita’s top technology trends, a recurring dis- cussion group that is a part of ala’s annual conference sand midwinter meetings. past top technology trends discussions can be found on lita’s blog (www.ala .org/ala/lita/litaresources/toptechtrends/toptechnol- ogy.cfm) and on lita’s website (www.ala.org/ala/lita/ litaresources/toptechtrends/toptechnology.cfm). the second source is the horizon project, a five-year qualita- tive research effort aimed at identifying and describing emerging technologies within the realm of teaching and learning. the project is a collaboration between the new media consortium and educause. the horizon project website (http://horizon.nmc.org/wiki/main_page) con- tains the annual horizon reports going back to . a final approach to project the future of it and libraries is to consider the work of our peers. the next library innova- tion may emerge from a sister institution. or perhaps it may take route at your local library first! reference . bejune, matthew m. and jana ronan. social software in libraries. arl spec kit . washington, d.c.: association of research libraries, . information technology and libraries | march currently we librarians seem to be hitching our wagon to the idea of library as community because in part it’s what we ourselves want. we’ve seen that our lita members want more community from our association, so it makes sense to us that our patrons also want community. it’s what pew, oclc, and other stud- ies seem to be telling us. the business-wired side of the world is breaking their backs to create every form of vir- tual community they can think of as quickly as possible. apply the appropriate amounts of marketing and then our patrons want those things and expect them from all of their historically important community resources, the library being a prime player in that group. so we strive and strive and strive to not only provide the standard issue face-to-face community we’ve always created, but to also create that new highly desired virtual community. either we create a library-specific version, or we at the very least create a way for our patrons to access those communities. hopefully, when our patrons step into those virtual communities, we work to make it possible for them to find libraries there, too. all well and good, but do we have a plan? what’s the goal? what’s the end achievement? if, as studies say, patrons with a research need turn to libraries first only one percent of the time, and instead first hit up friends and family fifty or more percent of the time, then where is our significance and place in either the physical or virtual spaces? we know we serve significant numbers in many ways. we have gate counts, circulation records, holds placed, warm bodies in the building—all manners of indi- cators that show a well-managed and -marketed library is in demand and appreciated. as we run into the terrible head-on crash of commu- nity and technology, willy-nilly doing absolutely every- thing we can to accommodate everyone and everything, because we’re librarians and library technologists and that’s what we do, do we really have a clue why we’re doing it? all fodder for deep thought and many lattes or beers and late night discussions. on the lita side, though, we’re embarking on doing something about this knot when it comes to serving our members. under the guidance of past-president bonnie postlethwaite we’ve established an assessment and research committee co-chaired by bonnie and diane bisom. to kick off the committee activities and to help them establish an agenda and direction, lita hired the research firm the wedewer group to work with the lita board and the new committee. stay tuned for reports and announcements from this committee as it works to find answers to some of those questions. and have that latte with a lita colleague as you seek to find some answers yourself. it’s all part of building community. mark beatty (mbeatty@wils.wisc.edu) is lita president / and trainer, wisconsin library services, madison. president’s message: doing something about life’s persistent problems? mark beatty article title | author editorial | truitt editorial: beginnings marc truitt as i write these lines in late february, the first hints of spring on the alberta prairie are manifest. alternatively, perhaps it’s just that the longer and warmer days are causing me to “think spring.” there are no signs yet of early bulbs—at least, none that i can detect with around a foot of snow in most places—but the sun is now rising at : a.m. and not setting until p.m., a dramatic change from the barely seven hours of daylight typical of december and january. and while none but the hardiest souls are yet outside in shorts and shirt-sleeves, somehow, daytime highs that hover around freezing seem downright pleasant in comparison with the minus thirties (not counting the wind chill) we were experiencing even a couple of weeks ago. yes, spring is in the air, even if the calendar says it is still nearly a month away. . . . so what, you may fairly ask, does the weather in edmonton have to do with ital? this is my first issue of ital as editor, and it may not surprise you to hear that i’ve been thinking quite a bit about what might be the right theme and tone for my first column. while i’ve been associated with the journal for quite awhile—first as a board member, and more recently as managing editor—my role has always been comfortably limited to background tasks such as refereeing papers and produc- tion issues. now, that is about to change; i am stepping a bit out of my comfort zone. it’s about beginnings. i follow with some awe in the footsteps of a long line of editors of ital (and jola, its predecessor). i’ve been honored to serve—and to learn a great deal—from the last two, dan marmion and john webb. you, the readers of ital, and i are fortunate to have as returning managing editor judith carter, who preceded me and taught me the skills required for that post; i hasten to emphasize that she is definitely not responsible for the things i did not do right in the job! regular readers of ital will recall that john webb often referred humorously and admiringly to the members of the ital editorial board as his “junkyard dogs;” he claimed that they kept him honest. with the addition of a couple of fine new members, i’m confident that they will continue to do so in my case! okay, with that as preface, enough about me . . . let’s talk about ital. ■ what’s inside this issue ital content has traditionally represented an eclectic blend of the best mainstream and leading/bleeding edge of library technology. we strive to be reflective of the broad, major issues of concern to all librarians, as well as alert to interesting applications that may be little more than a blip at the edge of our collective professional radar screen. our audience is not limited to those actively work- ing in library technology, although they certainly form ital’s core readership; rather, we seek to identify and publish content that will be relevant to all with an interest in or need to know about how technology is affecting our profession. thus, some articles will resonate with staff seeking new ways to use web . technologies to engage our readers, while other articles will be of interest to those interested in better exploiting the four decades’ worth of bibliographic metadata that forms the backbone of our integrated library systems. the current issue of ital is no exception in this regard. we lead off with two papers that reflect the renewed inter- est of the past several years in the role and improvement of the library online catalog. jia mi and cathy weng review opac interfaces, searching functionality, and results dis- plays to address the question of why the current opac is ineffective and what we can do to revitalize it. timothy dickey, in a contribution that received the lita/ exlibris student writing award, summarizes the challenges and benefits of a frbr approach to current and “next-gen” library catalogs. interestingly, as will become clear at the end of this column, dickey’s is not the first prize-winning frbr study to appear in the pages of ital. online learning has long been a subject of interest both to librarians and to the education sector as a whole. whereas the focus of many previous studies has been on the techniques and efficacy of online learning systems, though, connie haley’s paper takes a rather different approach, describing and exploring factors that character- ize the preference of learners for online training, as com- pared with more traditional in-person techniques. in gary wan’s and zao liu’s investigation of content- based information retrieval (cbir) in digital libraries, the authors describe and argue for systems that will enable identification of images and audio clips by automated comparison against digital libraries of image and audio files. finally, wooseob jeong prototypes an innovative application for enhancing web access by the visually impaired. jeong’s application makes use of force feed- back, an inexpensive, proven technology drawn from the world of video gaming. ■ some ideas about where we are going a change of editorship is always one of those good oppor- tunities for thinking about how we might improve, or of marc truitt (marc.truitt@ualberta.ca) is associate director, bibliographic and information technology services, university of alberta libraries, edmonton, alberta, canada, and editor of ital. information technology and libraries | march different directions we might explore. with that in mind, here are a couple of things we’re either going to try, or that we’re considering: different voices. ital’s format has long included provision for two “opinion” columns, one by the editor, and another by the president of lita. from time to time, past editors have given over their columns for guest edi- torials. however, there are many other voices that could enrich ital’s pages, and the existing structure doesn’t really have a “place” for the regular airing of these voices. beginning with the june issue, ital will include a regular column contributed by members of the board, on a rotating basis. the column will be about any topic related to technology and libraries that is on the author’s mind. i’m thinking about how we might expand this to include a similar column contributed by ital readers. while such reader contributions may lack the currency of a weblog, i think that they would make for thought- provoking commentary. oh, and there’s that “currency thing.” in recent years, those of us who bring you ital have—as have those responsible for other ala publications—discussed at length the whole question of when and how to move to a sustainable model of electronic publishing that will address the needs of readers. this issue is of course espe- cially important in the case of a technology-focused jour- nal, where content tends to age rapidly. unfortunately, for various reasons, we’re not yet at the stage where we can go completely and solely electronic. a recent conversation with one board member, though, surfaced an idea that i think in the meantime has merit: essentially, we might create a preprint site for papers that have been accepted and edited for future publication in ital. we might call it something such as ital express, and its mission would be to get content awaiting publication out and accessible. is this a “done-deal”? no, at this stage, it’s just an intriguing idea, and i’d be interested in hearing your views about it . . . or anything else related to ital, for that matter. you can e-mail me at marc.truitt@ualberta.ca. ■ and finally, congratulations dept. last week, martha yee, of the film and television archive at the university of california, los angeles received the alcts cataloging and classification section’s margaret mann citation for . martha was “recognized for her outstanding contributions to the practice of cataloging and her interest in cataloging education . . . [and her] professional contributions[, which] have included active participation in ala and alcts and numerous publica- tions.” of particular note, the citation specifically singled out her work in the areas of “frbr, opac displays, shared cataloging and other important issues, [in which] yee is making a significant contribution to the discussions that are leading the development of our field.” surely among the most important of these is her paper “frbrization: a method for turning online public finding lists into online public catalogs,” which appeared in the june issue of ital (p. – ). archived at the ital site, d-list, the cdl e-scholarship repository, and elsewhere, this seminal contribution has become one of the most accessed and cited works on frbr. we at ital are proud to have provided the original venue for this paper and congratulate martha on being named recipient of the margaret mann award. article title | author frbrization of a library catalog | dickey the functional requirements for bibliographic records (frbr)’s hierarchical system defines families of biblio- graphic relationship between records and collocates them better than most extant bibliographic systems. certain library materials (especially audio-visual formats) pose notable challenges to search and retrieval; the first benefits of a frbrized system would be felt in music libraries, but research already has proven its advantages for fine arts, theology, and literature—the bulk of the non-science, technology, and mathematics collections. this report will summarize the benefits of frbr to next- generation library catalogs and opacs, and will review the handful of ils and catalog systems currently operat- ing with its theoretical structure. editor’s note: this article is the winner of the lita/ ex libris writing award, . t he following review addresses the challenges and benefits of a next-generation online public access catalog (opac) according to the functional requirements for bibliographic records (frbr). after a brief recapitulation of the challenges posed by certain library materials—specifically, but not limited to, audio- visual materials—this report will present frbr’s benefits as a means of organizing the database and public search results from an opac. frbr’s hierarchical system of records defines families of bibliographic relationship between records and collocates them better than most extant bibliographic systems; it thus affords both library users and staff a more streamlined navigation between related items in different materials formats and among editions and adaptations of a work. in the eight years since the frbr report’s publication, a handful of working systems have been developed. the first benefits of such a system to an average academic library system would be felt in a branch music library, but research already has proven its advantages for fine arts, theology, and literature—the bulk of the non-science, technology, and mathematics collections. ■ current search and retrieval challenges the difficulties faced first, but not exclusively, by music users of most integrated library systems fall into two related categories: issues of materials formats, and issues of cataloging, indexing, and marc record structure. music libraries must collect, catalog, and support materi- als in more formats than anyone else; this makes their experience of the most common ils modules—circu- lation, reserves, and acquisitions—by definition more complicated. the study of music continues to rely on the interrelated use of three distinct information formats—scores (the notated manifestation of a composer’s or improviser’s thought), recordings (realizations in sound, and some- times video, of such compositions and improvisations), and books and journals (intellectual thought regard- ing such compositions and improvisations)—music libraries continue to require . . . collections that inte- grate [emphasis mine] these three information formats appropriately. put a different way, “relatedness is a pervasive char- acteristic of music materials.” this is why frbr’s model of bibliographic relationships offers benefits that will first impact the music collection. at present, however, musical formats pose search and retrieval challenges for most ils users, and the problem is certainly replicated with microforms and video recordings. the marc codes distinguish between material formats, but they support only one category for sound recordings, lumping together cd, dvd audio, cassette tape, reel-to- reel tape, and all other types. this single “sound record- ing” definition is easily reflected in opacs (such as those powered by innovative interfaces’ millennium and ex libris’ aleph ) and union catalogs (such as worldcat. org). however, the distinction between sound recording formats is embedded in subfields of the field, which presently cannot be indexed by many library automation systems because the subfields are not adjacent. an even more central challenge derives from the fact that music sound recordings—such as journals and essay collections—contain within each item more than one work. thus, for one of the central material formats collected by a music library (as well as by a public library or other aca- demic branches), users routinely find themselves searching for a distinct subset of the item record. perversely, though music catalogers do tend to include analytic added-entries for the subparts of a cd recording or printed score, and major ils vendors are learning to index them, aacr guidelines set arbitrary cutoff points of about fifteen tracks on a sound recording, and three performable units within a score. subsets of essay collections and journal runs are routinely exposed to users’ searches by indexing and abstracting services and major databases, but subsets of libraries’ music collections depend upon catalogers to exploit the marc records for user access. timothy j. dickey (dickeyt@oclc.org) is a post-doctoral researcher, oclc office of programs and research, dublin, ohio. frbrization of a library catalog: better collocation of records, leading to enhanced search, retrieval, and display timothy j. dickey information technology and libraries | march information technology and libraries | march in light of these pervasive bibliographic relation- ships, catalogers of music (again, with parallels in other subjects) have developed a distinctive approach to the marc metadata schema. in particular, they—with their colleagues in literature, fine arts, and theology—rely upon the t field for uniform work titles, and upon careful authority control. however, once again, many major ils portals have spotty records in affording access to library collections via these data. innovative interfaces’ millennium, though it clearly leads other major library products in this market, frequently frustrates music librarians (it is, of course, not alone in doing so). its automatic authority control feature works poorly with (necessary) music authority records. and even though innovative has been one of the first vendors to add a database index to the t field, partly in response to con- cerns expressed to the company by the music librarians’ user group, millennium apparently does not allow for an appropriate level of follow-through on searching. an initial search by name of a major composer, for instance, yields a huge and cluttered result set contain- ing all indexed t fields. the results do helpfully include the appropriate see also references, but those references disappear in a subsidiary (limited) search. in addition, the subsidiary display inexplicably changes to an unhelpful arrangement of generic fields (“mozart, symphonies”; “mozart, operas, excerpts”). similar chal- lenges will be faced by other parts of an academic or large public library collection, including the literature collec- tions (for works such as shakespeare’s plays), fine arts (for images and artists’ works), and theology (for works whose uniform title is in latin). the opac interfaces of other major ils vendors fare little better. the same search (for “mozart”) on the emory university library catalog (with an ils by sirsidynix), similarly yields a rich results set of more than one thousand records, and poses similar prob- lems in refining the search. in the case of this opac, an index of t fields also exists, but it only may be searched from the inside of a single record; as with millennium, sirsidynix’s interface will then group the next set of results confusingly by fields. the library corporation’s carl-x apparently does not contain a t index; the simple “mozart” search returns a much- simplified set of only results organized by a fields, and thus offers a more concise set of results but avoids the most incisive index for audio-visual materials. ex libris offers a somewhat more helpful display of its more restricted results; unfortunately for the present comparison, though the detailed results set does list the “format” of all mozart-authored items, the same term— “music”—is used for sound recordings, musical scores, and score excerpts, with no attempt logically to group the results around individual works. no t index appears present. ■ the frbr paradigm: review of literature and theory from the earliest library catalogs in the modern age, the tools of bibliographic organization have sought to afford users both access to the collection and collocation of related materials. anglo-american cataloging practice has traditionally served the first function by main entries and alternate access points and the second function by classification systems. however, as knowledge increases in scope and complexity, the systems of bibliographic control have needed to evolve. as early as the s, theo- ries were developing that sought to distinguish between the intellectual content of a work, and its often manifold physical embodiments. the paris international conference on cataloging principles first reified within the cataloging community a work-item distinction, though even the publication of the anglo-american cataloging rules, nd ed., “continued to demonstrate con- fusion about the nature . . . of works.” meanwhile, extensive research into the nature of bibliographic relationships groped toward a consen- sus definition of the entity-types that could encompass such relationships. ed o’neill and diane vizine-goetz examined some one hundred editions of smollett’s the expedition of humphrey clinker over a two-hundred-year span of publication history to propose a hierarchical set of definitions to define entity levels. the theoretical entities include the intellectual content of a work—which in the case of audio-visual works, may not even exist in any printed formats—the various versions, editions, and printings in which that intellectual content manifests itself, and the specific copies of each manifestation which a library may hold. research has discovered such clus- ters of bibliographically related entities for as much as percent or more of all the intellectual works in any given library catalog, and as many as percent of the works in a music catalog. this work laid the foundation for frbr (and, once again, incidentally underscored the breadth of its applicability to, and beyond, music catalogs). the theoretical framework of frbr is most concisely set forth in the final report of the ifla study group. the long-awaited publication traces its genesis to the stockholm seminar, and the resultant founding of the ilfa study group on functional requirements for bibliographic records. the study group set out to develop: a framework that identifies and clearly defines the entities of interest to users of bibliographic records, the attributes of each entity, and the types of relationships that operate between entities . . . a conceptual model that would serve as the basis for relating specific attri- butes and relationships . . . to the various tasks that users perform when consulting bibliographic records. article title | author frbrization of a library catalog | dickey the study makes no a priori assumptions about the bibliographic record itself, either in terms of content or structure. in other words, the intention of the group’s delibera- tions and the final report is to present a model for under- standing bibliographic entities and the relationships between them to support information organization tools. it specifically adopts an approach that defines classes of entities based upon how users, rather than catalogers, approach bibliographic records—or, by natural extension, any system of metadata. the frbr hierarchical entities comprise a fourfold set of definitions: ■ work: “a distinct intellectual or artistic creation”; ■ expression: “the intellectual or artistic realization of a work” in any combination of forms (including edi- tions, arrangements, adaptations, translations, per- formances, etc.); ■ manifestation: “the physical embodiment of an expression of a work”; and ■ item: “a single exemplar of a manifestation.” examples of these hierarchical levels abound in the bibliographic universe, but frequently music offers the quickest examples: ■ work: mozart’s die zauberflöte (the magic flute) ■ work: puccini’s la bohéme ■ expression: the composer’s complete musical score ( ) ■ manifestation: edition of the score printed by ricordi in ■ expression: an english language edition for piano and voices ■ expression: a performance by mirella freni, luciano pavarotti, and the berlin philharmonic orchestra (october ) ■ manifestation: a recording of this perfor mance released on ¹/³ rpm sound discs in by london records ■ manifestation: a re-release of the same per formance on compact disc in by london records ■ item: the copy of the compact disc held by the columbus metropolitan library ■ item: the copy of the compact disc held by the university of cincinnati in fact, lis research has tended to demonstrate what music librarians have always understood—that related- ness among items and complexity of families is most prevalent in audio-visual collections. even before the ifla report had been penned, sherry vellucci had set out the task: “to create new catalog structures that better serve the needs of the music user community, it is important first to understand the exact nature and complexity of the materials to be described in the catalog.” even limiting herself to musical scores alone (that is, no recordings or monographs), vellucci found that more than . percent of her sample exhibited at least one bibliographic relationship with another entity in the collection; she further related this finding to the very “inherent nature of music, which requires performance for its aural realization,” as opposed to, for example, monographic book printing. vellucci and others have frequently commented on how the relatedness of manifestations—in different formats, arrangements, and abridgements—of musical works continues to be a problem for information retrieval in the world of music bibliography. musical works have been variously and industriously described by musicologists and music bibliographers. yet, in the information retrieval domain [and, i might add, under both aacr and aacr ] . . . systems for bib- liographic information retrieval . . . have been designed with the document as the key entity, and works have been dismissed as too abstract . . . the work is the access point many users will bring—in their minds, and thus in their queries—to a system. they intend, however, to discover, identify, and obtain specific manifestations of that work. very recently, research has begun to demonstrate that the frbr model can offer spe- cific advantages to music retrieval in cases such as these: “the description of bibliographic data in a frbr-based database leads to less redundancy and a clearer presenta- tion of the relationships which are implicit in the tradi- tional databases found in libraries today.” explorations of the theory in view of the benefits to other disciplines, such as audio-visual and other graphic materials, maps, oral literature, and rare books, have appeared in the literature as well. the admitted weakness of the frbr theory, of course, is that it remains a theory at its incep- tion, with still preciously few working applications. ■ frbr applications working implementations of frbr to catalogs, opacs, and ilss are still relatively few but promise much for the future. the frbr theoretical framework has remained an area of intense research at oclc, which has even led to some prototype applications and, very recently, deploy- ment in the worldcat local interface. a scattered few other researchers have crafted frbr catalogs and catalog displays for their own ends; the library of congress has a prototype as well. innovative, the leading academic ils vendor, announced a frbr feature for release, information technology and libraries | march information technology and libraries | march yet shelved the project for lack of a beta-testing partner library. ex libris’ primo discovery tool, one other com- plete ils (by visionary technologies for library systems, or vtls), and the national library of australia, have each deployed operational frbr applications. the number of projects testifies to the high level of interest among the cataloging and information science commu- nities, while the relatively small number of successful applications testifies to the difficulties faced. oclc has engaged in a number of research projects and prototypes in order to explore ways that frbrization of bibliographic records could enhance information access. oclc research frequently notes the potential streamlining of library cataloging by frbrization; in addition they have experienced “superior presentation” and “more intuitive clustering” of search results when the model is incorporated into systems. work-level defini- tions stand behind such oclc research prototypes as audience level, dewey browser, fictionfinder, xisbn, and live search. in every case, researchers determined that, though it was very difficult to automate any identifi- cation of expressions, application of work-level categories both simplifies and improves search result sets. an algorithm common to several of these applications is freely available as an open source application, and now as a public interface option in oclc’s worldcat local. the algorithm creates an author/title key to cluster work- sets (often at a higher level than the frbr work, as in the case of the two distinct works that are the book and screenplay for gone with the wind). in the public search interface, the results sets may be grouped at the work level; users may then execute a more granular search for “all editions,” an option that then displays the group of expressions linked to the work record. unfortunately, as the software does not use t fields (its intention is to travel up the entity hierarchy, and it uses the xx, x, and fields), its usefulness in solving the above challenges may not be immediate. a somewhat similar application (though merrilee proffitt declares it not to be a frbr product) was redlightgreen, a user interface for the ex- rlg union catalog based upon quasi-frbr clustering. the reports from designers of other automated sys- tems offer interesting commentaries on the process. the team building an automatically frbrized database and user interface for austlit—a new union collection of australian literature among eight academic libraries and the national library of australia—acknowledged some difficulty with non-monographic works such as poems, though the majority of their database consisted of simpler work-manifestation pairs. based on strongly positive user feedback (“the presentation of information about related works [is] both useful and comprehensible”), a similar application was attempted on the australian national music gateway musicaustralia; it is unclear whether the project was shelved due to difficulties in automating the frbrization process. one recent application created for the perseus digital library adopts a somewhat different approach. rather than altering previously created marc records to allow hierarchical relationships to surface, this team created new records using crosswalks between marc and, for instance, mods, for work-level records. they claim some moderate level of success; though once again, their discussion of the process is more illuminating than their product. mimno and crane successfully allowed a single manifestation-level record to link upwards to many expressions, a necessary analytic feature especially for dealing with sound recordings. they did practically dem- onstrate the difficulty of searching elements from differ- ent levels of the hierarchy at the same time (such as work title and translator), a complication predicted by yee. three ils vendors have released products that use the frbr model: portia (visualcat), ex libris (primo), and vtls (virtua). the first product, a cataloging util- ity from a smaller player in the vendor market, claims to incorporate frbr into its metadata capture, yet the information available does not explain how, nor do they offer an opac to exploit it. the release of ex libris’ primo offers what the company calls “frbr groupings” of results. this discovery tool is not itself an ils, but promises to interoperate with major existing ils products to consolidate search results. it remains unclear at this time how ex libris’ “standard frbr algorithms” actu- ally group records; the single deployment in the danish royal library allows searching for more records with the same title, for instance, but does not distinguish between translations of the same work. vtls, on the other hand, has since offered a complete product that has the potential to modify existing marc records—via local linking tags in the and fields—to create frbr relationships. their own studies agreed with oclc that a subset, roughly percent, of existing catalog records (most heavily concentrated in music collections) would benefit from the process, and they thus allow for “mixed” catalogs, with only subsets (or even individually selected records) to be frbrized. the company’s own information suggests relatively sim- ple implementation by library catalogers, coupled with robust functionality for users, and may be the leading edge of the next generation of catalog products. ■ frbr solutions the ilfa study group, following its user-centered approach, set out a list of specific tasks that users of a computer-aided catalog should be able to accomplish: article title | author frbrization of a library catalog | dickey ■ to find all manifestations embodying certain criteria, or to find a specific manifestation given identifying information about it; ■ to identify a work, and to identify expressions and manifestations of that work; ■ to select among works, among expressions, and among manifestations; and ■ to obtain a particular manifestation once selected. it seems clear that the frbr model offers a framework of relationships that can aid each task. unfortunately, none of the currently available commercial solutions may be in themselves completely applicable for a single library. the oclc work-set algorithm is open source, as well as easily available through worldcat local, but it only works to create super-work records; it also ignores the t field so crucial to many of the issues noted above. none of the other home-grown applications may have code available to an institution. the virtua module from vtls offers a very tempting solution, but may require a change of vendor. either adapting one of these solutions or designing a local application, then, raises the question: what would the ideal system entail? catalog frbrization will tran- spire in two segments: enhancing the existing catalog to add bibliographic relationships to surface in the retrieval phase, and designing or adaptating a new interface and display to reflect the relationships. the first task may prove the more formidable, due to the size of even a mod- est catalog database and the difficulties often observed in automating such a task; while the librarians constructing the austlit system found a relatively high percentage of records could be transferred en masse, the oclc research team had difficulty automatically pinpointing expressions from current marc records. despite current technology trends toward users’ application of tags, reviews, and other metadata, a task as specialized as adding bibliographic relationships to the catalog demands specialized cataloging profession- als. the best approach within a current library structure may be to create a single new position to head the project and to act as liaison with cataloging staff in the vari- ous branches and with vendor staff, if applicable. each library branch may judge on its own the proportions of records to frbrize, beginning with high-traffic works and authors, those for whom search results tend to be the most overwhelming and confusing to users. each branch can be responsible for allocation of cataloging staff effort to the process, and will thus have specialist oversight of subsets of the database. three technical solutions to actually changing the database structure have been attempted in the literature to date: incrementally improving the existing marc records to better reflect bibliographic relationships, add- ing local linking tags, and simply creating new metadata schemas. the vtls solution of adding local linking tags seems most appropriate; relationships between records are created and maintained via unique identifiers and linking statements in the and fields. oclc’s open source software could expedite the creation of work-level records, and the creation of expression-level records will be made easier by the large amount of bib- liographic information already present in the current catalog. wherever possible, cataloging staff also should take the opportunity to verify or create links to authority files so as to enhance retrieval. creating a new catalog display option could be accom- plished via additions to current opac coding, either by adopting worldcat local or by designing parts of a new local interface. it need not even require a complete revi- sion; the single site (ucl) currently deploying vtls’ frbrized interface maintains a mixed catalog and offers, once again, a highly intuitive model. when a searcher comes across a bibliographic record for which frbr linking is available, they may click a link to open a new display screen. we should strive, however, to use simple interface statements such as “view all different kinds of holdings,” “this work has x editions, in y languages” or “this version of the work has been published z times” (both the oclc prototype and the austlit gateway offer such helpful and user-friendly statements). though the foundational work of both tillett and smiraglia focused upon taxonomies of relationships, the hierarchical struc- ture of the ifla proposal should remain at the forefront of the display, with a secondary organization by type of relationship or type of entity. rather than adopting a design which automatically refreshes at each click, a tree organization of the display should be more user-friendly, allowing users to maintain a visual sense of the organiza- tion that they are encountering (see appendix for screen- shots of this type of tree display). format information should be included in the display, as an indication of a users’ primary category, as well as a distinction among expressions of a work. with these changes, the library catalog will begin to afford its users better access to many of its core collec- tions. frbrization of even part of the catalog—concen- trating on high-incidence authors, as identified by subject specialists—will allow it better to reflect, and collocate, items within the families of bibliographic relationships that have been acknowledged a part of library collec- tions for decades. this increased collocation will begin to counteract the pitfalls of mere keyword searching on the part of users, especially in conjunction with renewed authority work. finally, frbr offers a display option in a revamped opac that is at the same time simpler than current result lists, and more elegant in its reflection of relatedness among items. each feature should better information technology and libraries | march information technology and libraries | march enable the users of our catalog to find, select, and obtain appropriate resources, and will bring our libraries into the next generation of cataloging practice. references and notes . ifla committee on the functional requirements for bib- liographic records, final report (munich: k. g. saur, ); see also http://www.ifla.org/vii/s /wgfrbr/bibliography.htm (accessed mar. , ). . this paper began as a graduate research assignment for lis (library automation), in the kent state university mlis program, march , . my thanks to jennifer ham- brick, nancy lensenmayer, and joan lippincott, for their helpful comments on earlier drafts. the curricular assignment asked for a library automation proposal in a specific library setting; the original review contained a set of recommendations concerning frbr through the lens of a (fictional) medium-sized academic library system, that of st. hildegard of bingen catholic univer- sity. as will be noted below, the branch music library typically serves a small population of music majors (graduate and under- graduate) within such an institution, but also a large portion of the student body that use the library’s collection to support their music coursework and arts distribution requirements. any music library’s proportion of the overall system’s holdings may be relatively small, but will include materials in a diverse set of formats: monographs, serials, musical scores, sound record- ings in several formats (cassette tapes, lps, cds, and stream- ing audio files), and a growing collection of video recordings, likewise in several formats (vhs, laser discs, and dvd). it thus offers an early test case for difficulties with an automated library system. . dan zager, “collection development and management,” notes—quarterly journal of the music library association , no. (march ): . . sherry l. velluci, “music metadata and authority con- trol in an international context,” notes—quarterly journal of the music library association , no. (mar. ): . . the opac for the university of huddersfield library system famously first deployed a search option for related items (“did you mean . . . ?”); http://www.hud.ac.uk/cls (accessed july , ). frbr not only offers the related item search, but also logically groups related works throughout the library cata- log. . allyson carlyle demonstrated empirically that users value an object’s format as one of the first distinguishing fea- tures: “user categorization of works: toward improved organi- zation of online catalog displays,” journal of documentation , no. (mar. ): – at . . millennium will feature heavily in the following discus- sion, both because of its position leading the academic library automation market (being adopted wholesale by, for instance, the ohio statewide academic library consortium), and because it was the subject of the original paper. . see alastair boyd, “the worst of both worlds: how old rules and new interfaces hinder access to music,” caml review , no. (nov. ), http://www.yorku.ca/caml/ review/ - /both_worlds.htm (accessed mar. , ); michael gorman and paul w. winkler, eds., anglo-american cataloging rules, nd ed. (chicago: ala, ). . in the past few years, a small subset of the search litera- ture has described technical efforts to develop search engines that can query by musical example; see j. stephen downie, “the scientific evaluation of music information retrieval systems: foundations and future,” computer music journal , no. (sum- mer ): – . a company called melodis corporation has recently announced a successful launch of a query-by-humming search engine, though a verdict from the music community remains out; http://www.midomi.com (accessed jan. , ). . see velluci, “music metadata and authority control in an international context”; richard p. smiraglia, “uniform titles for music: an exercise in collocating works,” cataloging and classification quarterly , no. ( ): – ; steven h. wright, “music librarianship at the turn of the century: technology,” notes—quarterly journal of the music library association , no. (mar. ): – . each author builds upon the foundational work of barbara tillett, “bibliographic relationships: toward a conceptual structure of bibliographic information used in cataloging” (ph.d. diss., university of california at los ange- les, ). . “at conferences, [my colleagues] are always groaning if they are a voyager client,” interview with an academic music librarian by the author, feb. , . . several prominent music librarians only discovered that innovative’s system had such a feature when instances of the automatic system’s changing carefully crafted music authority records were discovered; mark sharff (washington university in st. louis) and deborah pierce (university of washington), post- ings to innovative music users’ group electronic discussion list, oct. , , archive accessed feb. , . . music librarians are the only subset of the millennium users to have formed their own innovate users’ group. sirsi- dynix has a separate users’ group for stm librarians, and ex libris hosts a law librarians’ users’ group, two other groups whose interaction with the ils poses discipline-specific chal- lenges. . searches were tested on the the ohio state university libraries’ opac , http://library.osu.edu (accessed mar. , ). . http://www.emory.edu/libraries.cfm (accessed june , ). . searches performed on the library of oklahoma state university, http://www.library.okstate.edu (accessed june , ); tlc has considered making frbrization a possible fea- ture of their product. they offer some concatenation of “intel- lectually similar bibliographic records,” and “tlc continues to monitor emerging frbr standards”; don kaiser, personal communication to the author, july , . i was unable to reach representatives of sirsidynix on this issue. . searches performed on the mit library catalog, pow- ered by aleph http://libraries.mit.edu (accessed june , ). . eva verona, “literary unit versus bibliographic unit [ ],” in foundations of descriptive cataloging, ed. michael car- penter and elaine svenonius, – (littleton, colo.: libraries unlimited, ), and seymour lubetzky, principles of catalog- ing, final report phase i: descriptive cataloging (los angeles: institute for library research, ), are usually credited with article title | author frbrization of a library catalog | dickey the foundational work on such theories; see richard p. smira- glia, the nature of “a work”: implications for the organization of knowledge (lanham, md.: scarecrow, ), – , to whom the following overview is indebted. . anglo-american cataloging rules, cited in smiraglia, the nature of “a work,” . . among the many library and information science thinkers contributing to this body of research, the most prominent have been patrick wilson, “the second objective” in the conceptual foundations of descriptive cataloging, ed. elaine svenonius, – (san diego: academic publ., ); edward t. o’neill and diane vizine-goetz, “bibliographic relationships: implications for the function of the catalog,” in the conceptual foundations of descriptive cataloging, ed. elaine svenonius, – (san diego: academic publ., ); barbara ann tillett, “bibliographic relationships: toward a conceptual structure of bibliographic information used in cataloging” (ph.d. diss, university of california, los angeles, ); eadem, “bibliographic relation- ships,” in relationships in the organization of knowledge, carol a. bean and rebecca green, eds. , – (dordrecht: kluwer, ) (summary of her dissertation findings on – ); martha m. yee, “manifestations and near-equivalents: theory with special attention to moving-image materials,” library resources and technical services , no. ( ): – . . o’neill and vizine-goetz, “bibliographic relationships”; see also edward t. o’neill, “frbr: application of the entity- relationship model to humphrey clinker,” library resources and technical services , no. (oct. ): – . . theorists in music semiotics who have more or less pro- foundly influenced music librarians’ view of their materials include jean-jacques nattiez, music and discourse: toward a semi- ology of music, trans. by carolyn abbate (princeton, n.j.: princ- eton univ. pr., ), and lydia goehr, the imaginary museum of musical works (new york: oxford univ. pr., ). see also smiraglia, the nature of “a work,” . for a concise overview of how semiotic theory has influenced thinking about literary texts, see w. c. greetham, theories of the text (oxford: oxford univ. pr., ), – . . studies have found families of derivative bibliographic relationships in . percent of all worldcat records, . per- cent of records in the catalog of georgetown university library, . percent in the burke theological library (union theologi- cal seminary), . percent of theological works in the new york university library, and . percent in the sibley music library at the eastman school of music (university of rochester). see smiraglia, the nature of “a work,” , who cites richard p. smiraglia and gregory h. leazer, “derivative bibliographic relationships: the work relationship in a global bibliographic database,” journal of the american society for information science ( ): – ; richard p. smiraglia, “authority control and the extent of derivative bibliographic relationships” (ph.d. diss., university of chicago, ); richard p. smiraglia, “deriv- ative bibliographic relationships among theological works,” proceedings of the nd annual meeting of the american society for information science (medford, n.j.: information today, ): – ; and sherry l. vellucci, “bibliographic relationships among musical bibliographic entities: a conceptual analysis of music represented in a library catalog with a taxonomy of the relationships” (d.l.s. diss., columbia university, ). . ifla, final report, – . . ibid, – . . sherry l. vellucci, bibliographic relationships in music cata- logs (lanham, md.: scarecrow, ), . . ibid, ; . . vellucci, “music metadata”; richard p. smiraglia, “musi- cal works and information retrieval,” notes: quarterly journal of the music library association , no. (june ). patrick le boeuf notes that users of music collections often use the single word “score” to indicate any one of the four frbr entities; “musical works in the frbr model or ‘quasi la stessa cosa’: variations on a theme by umberto eco,” in functional require- ments for bibliographic records (frbr): hype or cure-all? ed. patrick le boeuf, – at – (new york: haworth, ). . smiraglia, “musical works and information retrieval,” . . marte brenne, “storage and retrieval of musical docu- ments in a frbr-based library catalogue” (masters’ thesis, oslo university college, ), . see also john anderies, “enhancing library catalogs for music,” paper presented at the conference on music and technology in the liberal arts environment, hamilton college, june , ; powerpoint presentation accessed mar. , , from http://academics. hamilton.edu/conferences/musicandtech/presentations/cata- log-enhancements.ppt; boyd, “the worst of both worlds.” . see the extensive bibliography compiled by ifla, cata- loging division: “frbr bibliography,” http://www.ifla.org/ vii/s /wgfrbr.bibliography.htm (accessed mar. , ). . the first ils deployment of the worldcat local applica- tion using frbr is with the university of washington libraries: http://www.lib.washington.edu (accessed june , ). . innovative interfaces, inc., “millennium preview: frbr support,” inn-touch (june ), . interestingly, the one- page advertisement for the new service chose a musical work, puccini’s opera la bohème, to illustrate how the sorting would work. innovative interfaces booth staff at the ala national conference, washington, d.c., june , , told the author the company has moved in a different development direction now (investing more heavily in faceted browsing). . denmark’s det kongelige bibliotek has been the first ex libris partner library to deploy primo, http://www.kb.dk/en (accessed july , ). the vtls system has been operating since at the université catholique de louvain, http:// www.bib.ucl.ac.be (accessed mar. , ). for austlit, see http://www.austlit.edu.au (accessed mar. , ). . rick bennett, brian f. lavoie, and edward t. o’neill, “the concept of a work in worldcat: an application of frbr,” library collections, acquisitions, and technical services , no. (spring ): – . work-level records allow manifestation and item records to inherit labor-intensive subject classification metadata; eric childress, “frbr and oclc research,” paper presented at the university of north carolina-chapel hill, apr. , , http://www.oclc.org/research/presentations/ childress/ -uncch-sils.ppt (accessed mar. , ). . thomas b. hickey, edward t. o’neill, and jenny toves, “experiments with the ifla functional requirements for bibliographic records (frbr),” d-lib , no. (sept. ), http://www.dlib.org/dlib/september /hickey/ hickey.html (accessed mar. , ). . thomas b. hickey and jenny toves, “frbr work-set algorithm,” apr. report, http://www.oclc.org/research/ projects/frbr/default.htm (accessed mar. , ); algorithm information technology and libraries | march information technology and libraries | march available at http://www.oclc.org/research/projects/frbr/algo- rithm.htm. on worldcat local, see above, note . . merrilee proffitt, “redlightgreen: frbr between a rock and a hard place,” http://www.ala.org/ala/alcts/alctsconted/ presentations/proffitt.pdf (accessed mar. , ). redlight green has been discontinued, and some of its technology incor- porated into worldcat local. . http://www.austlit.edu.au (accessed mar. , ), but unfortunately a subscription database at this time, and thus unavailable for operational comparison. see marie-louise ayres, “case studies in implementing functional requirements for bibliographic records: austlit and musicaustralia,” alj: the australian library journal , no. (feb. ): – , http:// www.nla.gov.au/nla/staffpaper/ /ayres .html (accessed mar. , ). . ibid. . see david mimno and gregory crane, “hierarchical catalog records: implementing a frbr catalog,” d-lib , no. (oct. ); http://www.dlib.org/dlib/october / crane/ crane.html (accessed mar. , ). . ibid. see also martha m. yee, “frbrization: a method for turning online public finding lists into online public cata- logs,” information technology and libraries , no. ( ): – , http://repositories.cdlib.org/postprints/ (accessed mar. , ). . portia, “visualcat overview,” http://www.portia.dk/ pubs/visualcat/present/visualcatoverview .pdf (accessed mar. , ); vtls, inc., “virtua,” http://www.vtls. com/brochures/virtua.pdf (accessed mar. , ). . http://www.exlibrisgroup.com/primo_orig.htm (accessed july , ). . syed ahmed, personal communication to the author, july , ; searches run july , , on http://www.kb.dk/en. the library’s holdings of manifestations of mozart’s singspiel opera, the magic flute, run to four different groupings on this catalog: one under the title “die zauberflöte,” one under the title “la flute enchantée: opéra fantastique en actes,” and two separate groups under the title “tryllefløtjen.” . “vtls announces first production use of frbr,” http:// www.vtls.com/corporate/releases/ / .shtml (accessed mar. , ). unfortunately, though this press release indi- cates commitments on the part of the université catholique de louvain and vaughan public libraries (ontario, canada) to use fully frbrized catalogs, only the first is operating in this mode as of july , and with only a subset of its catalog adapted. . virtua is not interoperable, for instance, with any of innovative’s other ils modules, which continue to dominate a number of larger academic consortia; john espley, vtls inc. director of design, personal communication to the author, mar. , . . see allyson carlyle, “fulfilling the second objective in the online catalog: schemes for organizing author and work records into usable displays,” library resources and technical services , no. ( ): – . . even at the work-level, yee distinguished fully eight dif- ferent places in a marc record in which the identity of a work may be located, “frbrization,” – . . gregory leazer and richard p. smiraglia imply that cataloger-based “maps” of bibliographic relationships are inad- equate; “bibliographic families in the library catalog: a quali- tative analysis and grounded theory,” library resources and technical services , no. ( ): – . the cataloging fail- ures they describe, however, are more a result of inadequacies in the current rules and practice, and do not really prove that catalogers have failed in the task of creating useful systems. . vinood chacra and john espley, “differentiating librar- ies though enriched user searching: frbr as the next dimensions in meaningful information retrieval,” powerpoint presentation, http://www.vtls.com/corporate/frbr.shtml (accessed mar. , ). . see yee, “frbrization.” . http://www.bib.ucl.ac.be (accessed mar. , ). . not only does the ex libris primo application need click- throughs, it creates a new window for an extra step before pre- senting a new group of records. bibliography anderies, john. “enhancing library catalogs for music.” paper presented at the conference on music and technology in the liberal arts environment, hamilton college, june , ; http://academics.hamilton.edu/conferences/musican- dtech/presentations/catalog-enhancements.ppt (accessed mar. , ). ayres, marie-louise. “case studies in implementing functional requirements for bibliographic records: austlit and musi- caustralia.” alj: the australian library journal , no. (feb. ): – ; http://www.nla.gov.au/nla/staffpaper/ / ayres .html (accessed mar. , ). bennett, rick, brian f. lavoie, and edward t. o’neill. “the concept of a work in worldcat: an application of frbr.” library collections, acquisitions, and technical services , no. (spring ): – . boyd, alistair. “the worst of both worlds: how old rules and new interfaces hinder access to music.” caml review , no. (nov. ); http://www.yorku.ca/caml/review/ - / both_worlds.htm (accessed mar. , ). brenne, marte. “storage and retrieval of musical documents in a frbr-based library catalogue.” masters’ thesis, oslo university college, . carlyle, allyson. “fulfilling the second objective in the online catalog: schemes for organizing author and work records into usable displays,” library resources and technical services , no. ( ): – . ______. “user categorization of works: toward improved orga- nization of online catalog displays.” journal of documenta- tion , no. (mar. ): – chacra, vinood, and john espley. “differentiating libraries though enriched user searching: frbr as the next dimen- sions in meaningful information retrieval.” powerpoint presentation, http://www.vtls.com/corporate/frbr.shtml (accessed mar. , ). childress, eric. “frbr and oclc research.” paper pre- sented at the university of north carolina-chapel hill, apr. , ; http://www.oclc.org/research/presentations/ childress/ -uncch-sils.ppt (accessed mar. , ). hickey, thomas b., and edward o’neill. “frbrizing oclc’s worldcat.” in functional requirements for bibliographic records article title | author frbrization of a library catalog | dickey (frbr): hype or cure-all? ed. patrick le boeuf, - . new york: haworth, . hickey, thomas b., and jenny toves. “frbr work-set algo- rithm.” apr. report; http://www.oclc.org/research/ frbr (accessed mar. , ). hickey, thomas b., edward t. o’neill, and jenny toves, “experiments with the ifla functional requirements for bibliographic records (frbr),” d-lib , no. (sept. ); http://www.dlib.org/dlib/september /hickey/ hickey. html (accessed mar. , ). ifla study group on the functional requirements for bib- liographic records. functional requirements for bibliographic records: final report. munich: k. g. saur, . layne, sara shatford. “subject access to art images.” in intro- duction to art image access: issues, tools, standards, strategies, murtha baca, ed., – . los angeles: getty research insti- tute, . leazer, gregory, and richard p. smiraglia. “bibliographic fami- lies in the library catalog: a qualitative analysis and grounded theory.” library resources and technical services , no. ( ): – . le boeuf, patrick. “musical works in the frbr model or ‘quasi la stessa cosa’: variations on a theme by umberto eco.” in functional requirements for bibliographic records (frbr): hype or cure-all? patrick le boeuf, ed., – new york: haworth, . markey, karen. subject access to visual resources collections: a model for computer construction of thematic catalogs. new york: greenwood, . mimno, david, and gregory crane. “hierarchical catalog records: implementing a frbr catalog.” d-lib , no. (oct. ); http://www.dlib.org/dlib/october /crane/ crane. html (accessed mar. , ). o’neill, edward t. “frbr: application of the entity-relationship model to humphrey clinker.” library resources and technical services , no. (oct. ): – . o’neill, edward t., and diane vizine-goetz. “bibliographic relationships: implications for the function of the catalog.” in the conceptual foundations of descriptive cataloging. elaine svenonius, ed., – . san diego: academic publ., . proffitt, merrilee. “redlightgreen: frbr between a rock and a hard place.” paper presented at the ala annual con- ference, orlando, fla.; http://www.ala.org/ala/alcts/alcts- conted/presentations/proffitt.pdf (accessed mar. , ). smiraglia, richard p. bibliographic control of music, – . lanham, md.: scarecrow and music library association, . ______. “content metadata: an analysis of etruscan artifacts in a museum of archaeology.” cataloging and classification quarterly, , no. / ( ): – . ______. “musical works and information retrieval,” notes: quarterly journal of the music library association , no. (june ): – . ______. the nature of “a work”: implications for the organization of knowledge. lanham, md.: scarecrow, . ______. “uniform titles for music: an exercise in collocating works.” cataloging and classification quarterly , no. ( ): – . tillett, barbara ann. “bibliographic relationships.” in relation- ships in the organization of knowledge. carol a. bean and rebecca green, eds., – . dordrecht: kluwer, . vellucci, sherry l. bibliographic relationships in music catalogs. lanham, md.: scarecrow, . ______. “music metadata and authority control in an interna- tional context.” notes—quarterly journal of the music library association , no. (mar. ): – . wilson, patrick. “the second objective.” in the conceptual foun- dations of descriptive cataloging. elaine svenonius, ed., – . san diego: academic publ., . wright, h. s. “music librarianship at the turn of the century: technology.” notes: quarterly journal of the music library association , no. (mar. ): – . yee, martha m. “frbrization: a method for turning online public finding lists into online public catalogs.” information technology and libraries , no. ( ): – ; http://reposi- tories.cdlib.org/postprints/ (accessed mar. , ). ______. “manifestations and near-equivalents: theory with spe- cial attention to moving-image materials.” library resources and technical services , no. ( ): – . zager, daniel. “collection development and management.” notes: quarterly journal of the music library association , no. ( ): – . information technology and libraries | march information technology and libraries | march a search on also sprach zarathustra on the online public access catalog for the universite catholique de louvain, with results frbrized. (a vtls opac). selecting the first work yields the following screen: . . . which, when frbrized, yields a list of expressions. any part of the tree may be expanded, to display manifes- tations, and item-level records follow. appendix: examples of a frbrized tree display article title | author online workplace training in libraries | haley this study was designed to explore and describe the relationships between preference for online training and traditional face-to-face training. included were variables of race, gender, age, education, experience of library employees, training providers, training locations, and institutional professional development policies, etc. in the library context. the author used a bivariate test, kruskal- wallis test and mann-whitney u test to examine the relationship between preference for online training and related variables. i n the era of information explosion, the nature of library and information services makes library staff update their work knowledge and skills regularly. workplace training has played an important role in the acquisition of knowledge and skills required to keep up with this information explosion. as richard a. swanson states, human resource development (hrd) is personnel training and development and organization develop- ment to improve processes and enhance the learning and performance of individuals, organizations, com- munities, and society (swanson ). training is the largest component of hrd. it helps library employees acquire more skills through continuous learning. online workplace training is a relatively new medium of deliv- ery. this new form of training has been explored in the literature of human resources development in corpora- tion settings (macpherson, elliot, harris, and homan ), but it has not been adequately explored in univer- sity and library settings. universities are unique settings in which to study hrd, and libraries are unique settings in which to examine hrd theory and practice. in human resource development literature there are studies on participation (wang and wang ) from the perspec- tive of individual motivation, attitudes, etc.; however, more research needs to be conducted to explore library employees’ demographics related to online training in the unique library contexts, such as various staff training and development, as well as training policies. hrd literature includes studies of online learning in formal educational settings (hiltz and goldman ; shank and sitze ; waterhouse ), and there are studies on relationships between national culture and the utility of online training (downey, wentling, wentling, and wadsworth ). but there has been very little research conducted in terms of online workplace training for library staff. it is not clear what relation- ships exist among preferences for online training and demographic variables such as ethnicity, gender, age, educational level, and years of library experience. due to lack of research in these areas, workplace training in libraries will be less effective if certain ethnic groups, or certain age groups, prefer traditional face-to-face train- ing as libraries move toward online training. the author believes that research should govern library practice. therefore, it is necessary to research this topic and dis- seminate the findings. because of the growth in online training, there is a need to gain a better understanding of these relationships. ■ purpose of the study the study aims to reveal the relationships between preferences for online or traditional face-to-face train- ing and variables such as ethnicity, gender, age, educa- tional level, and years of experience. it also studies the relationships among preference for online training and other variables of training locations, training providers, training budgets, and professional development policies. the constructs are: the preference for online training was related to demographics, library’s training budget, pro- fessional development policies, training providers, and the training locations. these factors were included in the research questionnaire. we begin with the research ques- tions, review the current literature, and then discuss the method, results, and need for further research. correlational research questions . what is the relationship between ethnicity and online workplace training preferences? . what is the relationship of employees’ educational levels, age, and years of library experience to online workplace training preferences? . how does preference for online workplace training in libraries relate to employee gender? . how does preference for online workplace training in libraries relate to training locations, training pro- viders, training budgets, and professional develop- ment policies? . do library staff prefer traditional face-to-face train- ing over online training? ■ review of the literature as stated above, training is the largest component of hrd. the discipline of hrd relies on three core theo- ries: psychological theory, economic theory, and system theory. swanson ( ) stated: online workplace training in libraries connie k. haley connie k. haley (chaley@csu.edu) is systems librarian, chicago state university library, illinois information technology and libraries | march information technology and libraries | march economic theory is recognized as a primary driver and survival metric of organizations; system theory recognizes purpose, pieces, and relationships that can maximize or strangle systems and subsystems; and psychological theory acknowledges human beings as brokers of productivity and renewal along with the cul- tural and behavioral nuances. each of these three theo- ries is unique, complementary, and robust. together they make up the core theory underlying the discipline of hrd (p. – ). three specific economic theory perspectives are believed to be most appropriate to the discipline of hrd: ( ) scarce resource theory, ( ) sustainable resource theory, and ( ) human capital theory (swanson ). training is an investment to human capital with valuable returns, but no costs. wenger and snyder’s study (as cited in mahmood, ahmad, samah, and idris ) states that today’s economy runs on knowledge and skills. thurow’s study (as cited in swanson ) states that new industries of the future depend on brain power. man-made com- petitive advantages replace the comparative advantage of natural-resources endowments or capital endowments. in a rapidly changing society, maintaining organiza- tional and individual competence has become a greater challenge than ever before (hake ). competences include knowledge, skills, and attitudes. much of the literature focuses on job-related functional competences (deist and winterton ). library workplace training is one of the primary methods of investing in human capital and increasing competence for library employ- ees. training is the process through which skills are developed, information is provided, and attributes are nurtured (davis and davis ). to increase training participation and efficacy, libraries need to determine employees’ preferences for online training or traditional face-to-face training; a resulting high training participa- tion rate would increase the competence of all employees. library trainers and administrators can encourage non- participants to attend training by offering different train- ing sessions (online or face-to-face), and/or by changing training policies and budget allocations. unlike person- ality and intelligence, skill competence may be learned; hence it may be improved through training and devel- opment (mcclelland ). nadler and tushman ( ) emphasized core competence as a key organizational resource that could be exploited to gain competitive advantage. core competence was defined as collective learning in the organization, especially how to coordinate diverse production skills and integrate multiple streams of technologies (prahalad and hamel ). mezirow ( ) asserted that there are asymmetrical power rela- tionships that influence the learning process (as cited in baumgartner ). learning more about the relation- ships may benefit training and learning. in other words, training may be more effective if it is provided in the form preferred by the majority of staff. as stated above, there is very little research about online workplace training for library staff. past stud- ies have focused on how to conduct online training for working catalogers (ferris ) or on online teaching for students (crichton and labonte ; hitch and hirsch ). from the design and implementation perspectives, kovacs ( ) discussed web-based training in libraries, and unruh ( ) emphasized problems in delivery of web-based training. markless ( ) addressed learning theory and other relevant theories that could be used to teach in libraries. yet there is a lack of research on the demographics of library staff participation in workplace training and a lack of research on the training preferences of library staff. ■ methodology the study took place in an online environment. the research activities covered a twenty-day period from april to april , . survey questionnaires and consent forms were posted on the web. select participants the survey url (http://freeonlinesurveys.com/render- survey.asp?id= ) was sent to library staff via library discussion lists along with a consent form including con- tact information and a brief explanation of the survey’s purpose. the surveys were anonymous and confidential. names, e-mail addresses, and personally identifiable information were not tracked. all participants filled out the survey online. the sample was limited to employ- ees who were at least nineteen years old. directors and department heads were also welcome to participate. instrument data collected for this study included categorical data (i.e., gender and ethnicity) and numeric data (age, years of education, and years of experience). this was an atti- tudinal survey; hence, the rensis likert scale was used for data feedback. most of the data was quantitative likert scale, such as the preference for online training, the professional development policy, and the budget allocation for training. data collection “entailed measur- ing the attitudes of employees, providing feedback to participants, and stimulating joint planning for improve- ment” (swanson ). likert-type scales provide more variation of responses and lend themselves to stronger statistical analysis (creswell ). it is important to select a well-tested instrument that article title | author online workplace training in libraries | haley reports reliable and valid data. however, measuring atti- tudes has been one of the most challenging forms of psy- chometric measurement (thorkildsen ). due to a lack of similar studies of libraries’ online training, no instru- ments could be found for this study except the education participation scale (eps), the deterrents to participation scale (dps), and the style analysis survey (sas) instru- ments. boshier’s forty-item eps ( ) is reliable in differ- entiating among diverse groups with varying reasons for participating in continuing education (as cited in merriam and caffarella ). the eps is used to find the motiva- tions as to why people participate in continuing education; consequently, the eps cannot answer all questions of this study. similarly, the dps reveals factors of nonparticipa- tion; hence, the dps cannot be used in this study. and while the sas is designed to identify how individuals pre- fer to learn, concentrate, and perform in both educational and work environments (sloan, daane, and giesen ), after careful examination, it was found that the sas was not well-suited to this study. because surveys are used to collect data and to assess opinions and attitudes (creswell ), the researcher chose to develop a survey that con- tained about items to assess library staff’s opinions and attitudes toward online training. the survey consisted of three parts: demographic variables, likert-scale assessment of online workplace training preference, and open-ended questions that were worded to reflect reasons for training preference (see appendix). to capture demographic data, participants were asked to indicate their age, years of library experience, years of education (high school/ged = ; two years college = ; bachelor’s degree = ; one master’s degree = ; two master’s degrees = ; ph.d/ed.d = +), gender ( = male or = female), ethnicity ( = asian/pacific islander, = american indian, = african american, = hispanic, = white, non-hispanic, and = other). the likert scale items are designed using a forced-choice likert scale (smith ), that is, an even number of response options ( = strongly agree; = agree; = mildly agree; = mildly disagree; = disagree; = strongly disagree), rather than an odd number (strongly agree; agree; neither agree nor disagree; disagree; strongly disagree). a scoring decision is consistently applied in order to have a meaningful inter- pretation of the scores. thus, for the likert scale items, the scaling method is to use high scores to represent stronger resistance to a measured attitude of online training. to insure reliability and validity of scores, the questionnaire was reviewed by an expert in the library field to validate if questions were representative of the library field. data collection the way a researcher plans to draw a sample is related to the best way to collect data (fowler ). the above sampling approach made it easier for data collection. the author collected data via the web survey company by paying for survey services on a monthly basis. the data was collected by the end of april . the total number of participants was (n= ), of which were valid. thirty-two participants did not com- plete the survey; those surveys with missing data were excluded from analysis. survey results were saved in a text file and then downloaded into spss for analysis. ■ results and analysis beside general frequency analysis, the kruskal-wallis test was used for six ethnic groups. since some ethnic groups had small sample sizes, all minorities ( ) were merged in one ethnic group. thus, the mann-whitney test was used for the two ethnic groups—minority and majority. the author also assessed bivariate relationships with preference of online training and other variables. frequencies analysis frequencies analysis includes demographics, preference of online training versus face-to-face training, budget, and professional development policies. demographics. eighty-five percent of participants were female, percent were white, percent had one master ’s degree, and percent had two or more mas- ter ’s degrees. nearly percent were forty years old or older; percent were fifty years old or older. thirty-six percent had less than years of library experience (see table ). preference of online training versus face-to-face training. most participants ( . percent) reported that online training was less effective than traditional face-to-face training. generally speaking, fewer participants ( . percent) preferred online training: strongly agree ( . percent), agree ( . percent), and mildly agree ( . percent). more participants ( . percent) did not prefer online training: mildly disagree ( . percent), disagree ( . percent), and strongly disagree ( . percent). budget. fifty-five percent of participants somewhat agree their library allocates sufficient budget for training: strongly agree ( . percent), agree ( . percent), and mildly agree ( percent). professional development policies. sixty-eight percent of participants somewhat agree their libraries had good professional development policies: strongly agree ( . percent), agree ( percent), mildly agree ( . percent). table shows the frequencies of preference of online training, budget, and policy. information technology and libraries | march information technology and libraries | march kruskal-wallis test of ethnicity (α = . ) in the kruskal-wallis test for ethnicity, to match the total number of minorities, white people were randomly selected from . the test was not significantly different. in the kruskal-wallis test, chi-square is . (df = ) and asymptotic significance was . , which was greater than the criterion α = . . there was no difference in pref- erence for online training between ethnic groups. mann-whitney u test statistics of ethnicity (α = . ) the mann-whitney test of ethnicity was not significant. asymptotic significance is . (z = -. ), which was greater than the criterion α = . . there was no difference in preference for online training between the minorities group and the group of white/not hispanic. mann-whitney u test statistics of gender (α = . ) the mann-whitney test of gender was not significant. asymptotic significance was . (z = -. ), which was greater than the chosen α value (α = . ). there was no significant difference in preference for online training between males and females. bivariate analysis (α = . ) bivariate correlations were computed (see table ). preference for online training was not associated with age, years of education, years of library experience, sufficient training budget, or professional development policy. it makes sense to believe that traditional face-to-face training has better quality than online training. before the survey analysis, the author expected that younger employees would prefer online training and older ones would prefer traditional face-to-face training due to the older employ- ees’ reluctance to change. it was also expected that highly educated employees would prefer online training while less educated ones, with fewer online skills, would prefer traditional face-to-face training. another assumption was that employees with more library experience would prefer online training while less experienced ones would prefer traditional face-to-face training. the survey showed these assumptions were wrong. it was also assumed that an insufficient training bud- get might result in a preference for online training, since online training is more cost effective; and that good pro- fessional development policies might result in preference for traditional face-to-face training because it is of better quality than online training. the survey found these assumptions to be false. training budget and professional development policies were irrelevant to the preference for online training. however, it was not surprising to find that preference for online training was associated with training providers and training locations, as seen in table . ■ discussion the exploration of the relationships among these vari- ables revealed that the preference for online training was not related to demographics, budgets, or profes- sional development policies. however, the preference for online training did show a correlation to training providers and locations. it was surprising to discover table . demographic characteristics characteristics frequency n % gender male female ethnicity asian/pacific islander . american indian . african american . hispanic . white . age – . – . – . – + . missing . education less than years . – years/bachelor . – years/one master . – years/two masters . + years/doctorate . missing data . years of library experience less than years . – years . – years . more than years . missing data . article title | author online workplace training in libraries | haley the preference for online training was not associated with ethnicity, gender, age, education, or library experi- ence. it was interesting to note that training budgets and professional development policies were not related to the preference for online training. several study hypotheses were confirmed. library staff preferred traditional face-to-face training as opposed to online training. although one-third ( . percent) of participants preferred (including mildly agree, agree, and strongly agree) online training, only . percent of participants thought that online training was more effec- tive than traditional face-to-face training. on the other hand, the majority ( percent) preferred online training when the training was held out of state; . percent preferred online training when it was held in state. the study concluded that online training was preferred if the training locations required participants to travel great distances from the library. of the participants, . percent preferred online training when the training was provided by a vendor. some participants did not think face-to-face contact was important for vendor training. this finding suggests that online training is a better choice for vendor training. fifty-five percent preferred online training when it was provided by an association/organization. association/ organization trainers should consider a combination of online and traditional face-to-face training to meet the needs of the majority. online training can be provided for some specific tasks, and supplemented by face-to-face training for others. the following are survey summaries of key reasons to use online and traditional training, along with suggestions from the survey participants. the main reasons to use online training ■ flexible (allows more people from one worksite to participate) ■ saves time ■ eliminates travel cost ■ generally lower training costs ■ ease of access (able to have hands- on practice with a technology and software program, able to refer back to supplemental materials, able to obtain wider range of training, appropriate to give general over- views in preparation for more in- depth face-to-face training) ■ convenient (have some control over one’s time, attend training from the comfort of home or office rather than having to drive somewhere and sit through a presentation, fits easily into a busy schedule, and self-paced in asynchronous online training) the main reasons to use face-to-face training ■ questions and answers: able to ask questions and discuss answers, see immediate feedback, questions others are asking may include some that you didn’t think of, and problems solved directly ■ networking with peers: face-to-face training allows for serendipitous networking opportunities, you have the option of personal conversations with train- ers as well as social opportunities to meet other pro- fessionals, it is hard to meet people and make friends through an online training, get out of the library once in awhile, find out what experiences staff from other departments or libraries are having ■ better communication and interaction: have per- sonal interaction with instructors and participants, share ideas and experiences with others, enjoy dis- cussions and diversity of personal opinions that come from face-to-face training ■ learn efficiently and effectively: learn from others— not just the instructor, get more out of real training, easy to get disinterested if no face-to-face contact, learn better from an instructor ■ technology barrier: sometimes technology can get in the way of training, some online training was poorly designed, online classes took forever to load and two seconds to read the whole page. suggestions to improve library workplace training administrative support. the most important factor is hav- ing library administrators who support training and encourage staff at all levels to attend training. provide workshops for professional librarians and civil service workers that relate to their work, and give them release time for training. library administrators must under- stand the importance of training and develop training policies with a commitment toward staff development. table . frequencies of preference of online training, budget, and policy descriptor frequency mean* median* std. deviation preference of online training . . budget . . . policy . . . = strongly agree; = agree; = mildly agree; = mildly disagree; = disagree; = strongly disagree information technology and libraries | march information technology and libraries | march library administrators must plan and design train- ing infrastructures for core competence and cumulative learning, instead of spontaneous one-shot training for new products or systems. more training. many participants expressed their desire for more training. training not only increases their knowledge and skills needed for their job, but also provides opportunities to network with colleagues. more face-to-face and technical hands-on training are needed since many librarians felt left out of the technol- ogy loop. they think that maintaining a current view of developments in technology is difficult. more online training is needed, both asynchronous and synchronous. asynchronous training is good for self-paced training, which is preferred by many survey participants, while some enjoy online webcasts of seminars and workshops for better interaction. it is hoped that state libraries will provide online streaming videos about various topics for academic, private, and public library staff. more funding. make more funding available for library workplace training. the training budget should not be the first thing cut when budgets get tight. a combination of online and traditional face-to-face training. walton ( ) notes that we must ensure we learn and grow. we may learn and grow by participat- ing workplace training. training pro- grams should be built into strategic hrd plans that will best fit employ- ees’ learning preference. this study shows that online training works well with basic informational topics and most technology topics (databases, searching, or web-related technolo- gies). certain simple topics were more appropriate for online training, such as a vendor ’s product and procedural training. some topics do not translate well into online training, however, such as how to conduct storytimes— topics that require a lot of interaction between participants. difficult topics need traditional training for direct answers from the instructor. topics that need in-depth discussion should be provided with traditional training. in other words, provide basic train- ings online and save face-to-face train- ing for more difficult topics. ■ future research future research should focus on new learning needs, how people interact with technology, and how people learn in an online environment. more research is needed for a variety of online training. in this study, the generic term “online training” was used. future study needs to expand the term “online training” to static asynchronous online training and interactive synchronous online training. static online training includes text-only static, and text- graphic static with or without voice. interactive online training includes voice-only interactive, and voice-video interactive with ability to ask and answer questions in real time. as time goes by, more people will have taken online training and will be more comfortable with it. as more people have online training experience, their attitudes toward online training may change. further research should examine and measure library staff’s preference for a variety of online training. in addition, participants should be surveyed by grouping experienced online trainees and non-experienced online trainees. finally, studies may be conducted to survey library staff in other countries to com- pare their preferences with those of their u.s. peers. the goal of this study is to provide helpful infor- mation for department heads, supervisors, and library human resources staff to assist them in determining the types of training that will be most effective to meet train- table . bivariate correlations with preference of online training ( α = . ) variables preference of online training age . education . library experience . budget . prof. development policies . training provider vendors <. * associations/org. (ala, oclc, etc.) <. * lib. consortia <. * library/institution <. * training location out of state <. * in state <. * in town <. * in house <. * * significant at α = . article title | author online workplace training in libraries | haley ing needs. the author hopes this study also provides useful information to all library employees who attend training or workshops, including civil service person- nel and librarians, and that this study will be utilized for further research on library training and, in turn, that research will make more contributions to the workplace training literature of libraries and other professions. acknowledgements the author thanks lorraine lazouskas, john webb, judith carter, and the copy editors at ala production services for their assistance and valuable input on this manuscript. bibliography baumgartner, l. m. . preface. in l. m. baumgartner and s. b. merriam (eds.), adult learning and development: multicultural stories. malabar, fla.: krieger publishing. creswell, j. w. . educational research: planning, conducting, and evaluating quantitative and qualitative research. nd ed. upper saddle river, n.j.: pearson merrill prentice hall. crichton, s., and r. labonte. . innovative practices for inno- vators: walking the talk; online training for online teaching. educational technology & society , no. : – . davis, j. r., and a. b. davis. . effective training strategies : a comprehensive guide to maximizing learning in organizations. san francisco: berret-koehler. deist, f. d., and j. winterton. . what is competence? human resource development international , no. : – . downey, s., r. m. wentling, t. wentling, and a. wadsworth. . the relationship between national culture and the usa- bility of an e-learning system. human resource development international , no. : – . ferris, a. m. . cataloging internet resources using marc and aacr : online training for working catalogers. catalo- ging and classification quarterly , no. : – . fowler, f. j. . survey research methods. rd ed. thousand oaks, calif.: sage. haley, c. k. . who participates in online workplace training in libraries? survey results retrieved april , , from http:// freeonlinesurveys.com/viewresults.asp?surveyid= . hake, b. j. . lifelong learning in late modernity: the chal- lenges to society, organizations, and individuals. adult edu- cation quarterly , no. : – . hiltz, s. r., and r. goldman, eds. . learning together online. mahwah, n.j.: lawrence erlbaum associates. hitch, l. p., and d. hirsch. . model training. journal of aca- demic librarianship , no. : – . kovacs, d. k. . designing and implementing web-based training in libraries. business and finance division bulletin (winter): – . macpherson, a., m. elliot, i. harris, and g. homan. . e-lea- rning: reflections and evaluation of corporate programme. human resource development international , no. : – . mahmood, n. h. n., a. ahmad, b. a. samah, and k. idris. . informal learning of management knowledge and skills and transfer of learning among head nurses. in human resource development in asia: harmony and partnership, r. moon, a. m. osman-gani, k. shinil, g. roth, and h. oh, eds. seoul: the korea academy of hrd. markless, s. . learning about learning rather than about tea- ching. retrieved july , , from http://www.ifla.org/iv/ ifla /papers/ - e.pdf. mcclelland, d. . identifying competencies with behavioral- event interviews. psychological science , no. : – . merriam, s. b., and r. s. caffarella. . learning in adulthood. san francsico: jossey-bass. mezirow, j. . learning to think like an adult: transformation theory; core concepts. in learning as transformation: critical perspectives on a theory in progress, j. mezirow and associates, eds. san francisco: jossey-bass. nadler, d. a., and m. tushman. . the organization of the future: strategic imperatives and core competencies for the st century. organisational dynamincs , no. : – . prahalad, c. k., and g. hamel. . the core competence of the corporation. harvard business review , no. : – . shank, p., and a. sitze. . making sense of online learning. san francisco: pfeiffer. sloan, t., c. j. daane, and j. giesen. . mathematics anxiety and learning styles: what is the relationship in elementary preservice teachers? school science and mathematics , no. : – . smith, j. t. . applied categorical data analysis. lecture presented in spring at northern illinois university, dekalb. swanson, r. a. . foundations of human resource development. san francisco: berrett-koehler. thorkildsen, t. a. . fundamentals of measurement in applied research. boston: pearson education. unruh, d. l. . desktop videoconferencing: the promise and problems of delivery of web-based training. internet and higher education , no. : – . walton, j. . strategic human resource development. harlow, england: pearson education. wang, g. g., and j. wang. . toward a theory of human resource development learning participation. human resource development review , no. : – . waterhouse, s. . the power of elearning: the essential guide for teaching in the digital age. boston: pearson education. information technology and libraries | march information technology and libraries | march appendix. questionnaire part i. . gender q male q female . ethnicity q asian or pacific islander q american indian q african american q hispanic q white, non-hispanic q other ____ . please indicate the year of your birth: _________ . please indicate years of education: _________ . please indicate years of library experience: ________ part ii. for questions – , please read each item and check the response that best matches your degree of agreement/disagree- ment: ( = strongly agree; = agree; = mildly agree; = mildly disagree; = disagree; = strongly disagree) . if training is provided by library vendors such as ebsco or blackwell, i would prefer that it be offered online rather than face-to-face. . if training is provided by associations/organizations such as ala and oclc, i would prefer that it be offered online rather than face-to-face. . if training is provided by library consortia, i would prefer that it be offered online rather than face-to-face. . if training is provided by your institution or library, i would prefer that it be offered online rather than face-to- face. . if training location is out of state, i would prefer that it be offered online rather than face-to-face. . if training location is in-state, i would prefer that it be offered online rather than face-to-face. . if training location is in town, i would prefer that it be offered online rather than face-to-face. . if training location is in-house, i would prefer that it be offered online rather than face-to-face. . my library allocates sufficient budget for training (may include online training). . my library has good professional or staff development policies. . generally speaking, i prefer online training rather than face-to-face training. part iii. . state reasons for your preference of traditional face-to-face training. . state reasons for your preference of online training. . please make suggestions to improve library workplace training. . do you think that online training is less effective than traditional face-to-face training? yes__ no __ article title | author revitalizing the library opac | mi and weng the behavior of academic library users has drastically changed in recent years. internet search engines have become the preferred tool over the library online public access catalog (opac) for finding information. libraries are losing ground to online search engines. in this paper, two aspects of opac use are studied: ( ) the current opac interface and searching capabilities, and ( ) the opac bibliographic display. the purpose of the study is to find answers to the following questions: why is the current opac ineffective? what can libraries and librarians do to deliver an opac that is as good as search engines to better serve our users? revitalizing the library opac is one of the pressing issues that has to be accomplished. t he information-seeking behavior of today’s aca- demic library users has drastically changed in recent years. according to a survey conducted and published by oclc in , approximately per- cent of college students across all the regions that were included in the study (including areas outside the united states) begin their electronic information searches with internet search engines. more than half of u.s. residents used google for their searches. internet search engines dominate the information-seeking landscape. academic libraries are the ones affected most, because many col- lege students are satisfied with the answers they find on the internet for their assignments, and they end up not taking advantage of the many quality resources in their libraries. for many years, before the internet search engine emerged, library catalogs were the sole information-seek- ing gateway. just as the one-time industry giant kodak has lost ground to digital photography, academic library opacs are losing ground to online search engines. all along we academic librarians have devotedly and assidu- ously produced good cataloging records for the public to use. we have diligently and faithfully educated and helped our faculty and students find the proper library resources to fulfill their research needs and assignment requirements. we feel good about what we have achieved. why have our users switched to online search engines? ■ the evolution of user behavior it is technology and rising user expectations that have contributed to the changes in user behavior. as coyle and hillmann pointed out: “today’s library users have a different set of information skills from those of just a few decades ago. they live in a highly interactive, networked world and routinely turn to web search engines for their information needs.” a recent study conducted by the university of georgia on undergraduate research behav- ior in using the university’s electronic library concluded that internet sites and online instruction modules are the primary sources for their research. the students’ year of study did not make much of a difference in their choices. tenopir also concluded from her study of approximately scholarly works published between and that no matter what type of resources were used, “con- venience remains the single most important factor for information use.” recently, oclc identified three major trends in the needs of today’s information consumers—self-service (moving to self-sufficiency), satisfaction, and seamless- ness. services provided by google, amazon, and similar companies are the major cause of these emerging trends. customers have wholeheartedly embraced these prod- ucts because of their ease of use and quick delivery of “good enough” results. researchers do not need to take information literacy classes to learn how to use an online search engine. they do not need to worry about forget- ting important but infrequently used search rules or com- mands. in addition, the search results delivered by online search engines are sorted using relevance ranking sys- tems that are more user-friendly than the ones currently employed by academic library opacs. these are just some of the features that current academic library opacs fail to deliver. in , campbell and fast presented their analysis of an exploratory study of university students’ perceptions of searching opacs and web search engines. they found that “[s]tudents express a distinct preference for search engines over library catalogues, finding the catalogue baffling and difficult to use effectively.” as a result, library opacs, because they do not fulfill user needs, have been bombarded with criticism. we often hear librarians complain about how library users forget what they have learned in user education classes. librarians sometimes even laugh at users’ igno- rance and ineffectiveness in searching library opacs. this legacy mentality has actually prevented librarians from recognizing the changes in user behavior and expectations that have occurred in the past decade. rarely have librar- ians considered ineffective opac design to be at the root of unsuccessful opac use. roy tennant has mentioned frequently in his presentations that “only librarians like to search; users prefer to find”; that “users aren’t lazy, they are jia mi and cathy weng revitalizing the library opac: interface, searching, and display challenges jia mi (jmi@tcnj.edu) is electronic resources/serials librarian and cathy weng (weng@tcnj.edu) is head of cataloging, the college of new jersey library, ewing. information technology and libraries | march information technology and libraries | march human.” it is only natural that library users turn to internet search engines first for their information needs. ■ the opac reexamined cutter, in his book, introduced the objectives of the library catalog as follows: . to enable a person to find a book of which either a. the author b. the title is known c. the subject . to show what the library has a. by a given author b. on a given subject c. in a given kind of literature . to assist in the choice of a book a. as to its edition (bibliographically) b. as to its character (literary or topical) the majority of today’s opacs have successfully ful- filled cutter’s model in finding known items. following the card-catalog convention, bibliographic elements such as title, author, and subject have been the leading search options in opac search menus for many years. it was assumed that users always came to the library with spe- cific author, title, or subject information in mind before searching the catalog. the opac bibliographic display is in essence an electronic version of the card catalog. to accommodate the bibliographic data from card catalogs, many display labels were created, but often without regard to whether or not they were suitable in an online environment. this data-centered, card-catalog type of design was easily understood and fluently used by librar- ians, but not by most end users. campbell and fast found in their study that “while the participants were generally happy with their understanding of search engines, they frequently expressed a low opinion of their ability to search the catalogue.” they also found that students felt that “[t]he web is cluttered; the catalogue is organized. however, this organization was not always helpful; it was admired, but not understood.” the traditional catalog retrieval mechanism is sig- nificantly different from the web search engine. as yu and young noted in , “web search engines and online bookstores have a number of features that are not typically incorporated into opacs. these functions include: natural-language entry, automated mapping to controlled vocabulary, spell-checking, similar pages, rel- evance-ranked output, popularity tracking, and brows- ing.” these features have unquestionably affected user expectations in searching library opacs. teaching users to search for structured bibliographic data is completely opposed to the ever-popular free and open internet search mechanism drawn from the google-like search experience, which does not require any special training. since academic libraries aim to provide more dynamic and versatile services, revitalizing library opacs should be considered a top priority. furthermore, librarians’ expectations of user behavior should adjust to today’s needs. educating users to become fluent in using opac search commands and rules has become less relevant as users now seldom read and follow instructions. investing effort and energy in designing a truly user-friendly opac that functions intuitively to achieve productive retrieval could not be more imperative. academic librarians have started pondering what changes should be made to library opacs so that a truly user-friendly, twenty-first-century catalog that offers a “google-like” experience can be delivered. two impor- tant aspects that affect the usability of library opacs are addressed in this article: ( ) the current interface and searching capabilities and ( ) the bibliographic display. the opac’s public interface and searching capabilities together function as a finding aid. it determines how successful a user is in retrieving information and is the gateway to library resources. the effectiveness of an opac’s bibliographic display affects the user’s under- standing of the bibliographic description. users use bibliographic information to identify, select, and obtain library resources. ■ the study of the public interface of library opacs to find out how academic libraries designed and admin- istered their opacs, the authors examined the interfaces of association of research libraries (arl) libraries’ opacs powered by five major integrated library systems (ils): aleph, horizon, millennium, unicorn, and voyager. the study focused on searching ability, relevance rank- ing, layout, and linking functionalities. during the study, we expected each ils system to have its own opac design. we also anticipated that search mechanisms would be managed differently at each location. however, we were surprised by the great disparities that we discovered in opac quality, a clear indication of the time and effort (or lack thereof) devoted to their maintenance and improvement. the findings are summarized below. google-driven changes—keyword search as the default search key in his article “mental models for search are getting firmer,” usability expert jakob nielsen argued that cur- > article title | author revitalizing the library opac | mi and weng rent users have already developed a firm mental model of searching: search is such a prominent part of the web user experi- ence that users have developed a firm mental model for how it’s supposed to work. users expect search to have three components: ■ a box where they can type words ■ a button labeled “search” that they click to run the search ■ a list of top results that’s linear, prioritized, and appears on a new page—the search engine results page (serp) in our experience, when users see a fat “search” button, they’re likely to frantically look for “the box where i type my words.” the mental model is so strong that the label “search” equals keyword searching, not other types of search. studies have also shown that the default search option to which an opac is set affects users’ success in retriev- ing information. two studies on university opac search transactions confirmed that novice users preferred search- ing by keyword. at nanyang technological university, singapore, a recent search transaction log study was conducted to “identify query and search failure patterns with the goal of identifying areas of improvement for the system.” results indicated that “the most commonly used search option for the ntu opac is the keyword search. the use of keyword searches contributed to . percent of all queries while other options such as title, author, and subject accounted for . percent, . percent, and . percent of all searches respectively.” at california state university–los angeles, a four- quarter ( – ) search transaction log analysis also revealed similar results. after the library implemented an “advanced keyword search” feature that provided more user-centered, behind-the-scenes search algorithms and that set keyword search as the default, the keyword search queries rose dramatically. many university library opacs have already begun to adopt features employed by internet search engines. among the arl library opacs studied, have “keyword(s) anywhere” as the default search key (see appendix a). this is a positive sign that libraries are pay- ing attention to user search behavior. thirty-six libraries’ default search keys are still set to “title,” and six libraries, instead of providing a default search option, list field choices from which users must choose before entering their search terms. the title search used as the default option holds some potential problems. in order to retrieve good results from a title search, users are expected to type in a title in the right order, spelled correctly, and omitting the initial article (a, an, the), if any. while librarians are fluent with these seemingly simple rules, students and faculty con- stantly have trouble remembering them. providing online search tips and offering information literacy classes only help a little. since presenting keyword search as the default has proved effective, libraries using title search as their opac default search option might want to recon- sider switching their default setting to keyword. search ability—true keyword search the basis of current opac search systems is boolean logic. the ease of using google-like search engines comes from its implicit “and” feature, which eliminates the need to enter boolean connectors (and, or, not) between search terms. this is logical because users usually look for records that contain all the terms that they enter. sixty-six percent of the arl libraries studied have opacs with keyword set as the default search option. these libraries handle boolean logic in keyword search- ing very differently. all five ils vendors offer “automatic and” functionality, but not all of these libraries have adopted it: in some cases, users are required to enter boolean operators during a search. emory university library’s opac automatically executes “same” for mul- tiple search words if no boolean operators are entered which means that it will find records with the search terms in the same bibliographic fields. syracuse university’s opac automatically uses the boolean operator “or” for all keyword queries. this practice can generate too many irrelevant results. libraries that automatically supply the boolean operator “and” for multiple terms entered in the search box consequently produce more relevant results. in addition, none of the arl opacs studied han- dle auto-correction for typos, spell-check, auto-plurals, auto-word-truncation, punctuations, or special charac- ters. this makes searching unnecessarily inconvenient. for many years now, teaching students how to prop- erly use boolean operators has been one of the essen- tial topics in information literacy classes. after taking these classes, do students use boolean operators when searching? a study of , transaction logs collected by french universities revealed that french uni- versity students use boolean operators infrequently. fifty-six percent of the queries used only a single term. approximately percent of the queries contained one boolean operator. to further investigate the impact of information search expertise on the use of boolean opera- tors, the study showed that approximately one-third ( percent) of the students (considered the “novice” group in the study) still did not use boolean operators even when they were explicitly invited to do so, compared to percent of librarians (considered the “expert” group in the study), who used at least one boolean operator for their queries. therefore, complicated search strategies and syntax are mostly used by expert users. novice users information technology and libraries | march information technology and libraries | march prefer to use natural-language queries. libraries also handle phrase searching in different ways. phrase searching usually is embedded within keyword search either explicitly or implicitly depending upon the ils system. aleph (ex libris) libraries use a radio button for “word or phrase” or “words adjacent” or “exact phrase” options for the computer to execute the command. unicorn (sirsi/dynix) libraries provide three options: “keyword,” “begins with,” and “exact.” some libraries have the “exact” command executed to search every field in a bibliographic record; other librar- ies search the title, subject, and author fields only. the millennium system’s (innovative) keyword search fea- ture can do automatic phrase and “and” search. some millennium libraries (e.g., michigan state university) take advantage of this feature to search words entered as phrases first and, if unsuccessful, the system then repeats the search for the same words using the boolean opera- tor “and.” this feature produces more relevant search results. however, several millennium libraries have not implemented this feature. they still use “boolean key- word” search as the default and instruct users to add quotation marks to define phrases. the voyager (ex libris, formerly endeavor) system offers two types of keyword searches: “keyword relevance” and “keyword boolean.” both options can handle phrase searching. but users are required to enter quotation marks for specific terms used as phrases. some libraries intentionally made only one keyword search option available. other librar- ies provided both options and used different languages as an opac search key (see appendix b). these search keys are not self-explanatory, and users will often find them puzzling. the default help screen provided by the ils vendor and adopted by many voyager libraries does not help much either (see appendix c). thirty-one of the voyager libraries provide a boolean keyword search option. only five libraries utilize the automatic “and” feature. one library uses boolean keyword search as the only keyword option, but did not activate the automatic “and” functionality. relevance ranking in search results when users search by keyword, the best way to sort the results is by relevance. presenting the most relevant results at the top of the results page is crucial because it enhances library resource discovery and access. other sorting options, such as title or publication date, are not very useful since users usually do not have titles or pub- lication dates in mind when browsing search results from a keyword search. three ils systems (millennium, unicorn, and voyager) have a relevance-ranking feature, yet this functionality was very much underutilized by the libraries studied. of the eighteen unicorn libraries, only five offered relevance ranking. none made it the default sorting option. thirty- six of the millennium libraries provided relevance ranking as a sorting option. only twelve of those librar- ies made relevance ranking the default sorting system. twenty-seven out of the thirty-five voyager libraries offered the keyword (relevance) search option, under which the search results were automatically ranked by relevance. out of the twenty-nine voyager libraries that offered the keyword (boolean) search option, only four libraries used relevance as the default sorting system. the rest of the libraries used a “system sort” mechanism that sorted search results by bibliographic control num- ber. figure summarizes the sorting options used by the arl libraries studied and also shows the default sorting options for keyword search. unlike online search engines, which pull data directly from full-text documents, library opacs search for words from the structured metadata entered by catalog- ers. different fields are set to carry different weights for relevance considerations. the behind-the-scenes algo- rithm (the criteria used to decide the level of relevance) should be carefully established to warrant a good ranking scheme. for example, the new opac of north carolina state university library, powered by endeca, adopted an algorithm based on field weighting, phrase matching, facet lcsh, term frequency (tf), and inverse document frequency (idf). their search results are indeed more logically ranked by relevance. recently there have been suggestions to incorporate circulation statistics, book review data, and a library of congress call number table into the algorithm. the checkout data would provide a rough substitute for google’s pagerank (a count of links to a site, which is an indication of the site’s popularity), and book reviews would provide more text to be consid- ered in the relevancy tests. using library of congress call numbers would either require having the call number table loaded and then running the search terms against it or including call numbers in the algorithm, giving more weight to titles having the same call number. for example, seven out of twenty-three results generated for a search for “new york history” on an opac have the call number “f .” the call number “f ” is linked to the call number table with the subject new york and history. it can be confirmed that seven items with call number “f ” should be considered more relevant and ranked first on the results list. more research needs to be done in this area. the search results display the search results display is critical. the information, options, and bibliographic data presented on the browse page help users decide what actions to take next. in the opacs examined, the authors found the following problems: article title | author revitalizing the library opac | mi and weng . search terms and search boxes were not retained on the results page after a search is performed, many opacs do not effectively carry the original search information onto the results screen. this information includes the search key and the words typed in the search box. users need to con- sult this information to identify and select records relevant to their needs from the search results page. based on the retained information, users also decide what to do next. for example, they might change their search strategy or modify their previous search. many of the opacs studied neglected to display the original search information. even better than just displaying the text of the user’s search terms would be to maintain them in search boxes at the top and bottom of the results display page. this way, users would only have to modify their search terms rather than type new search terms each time they wished to modify their original search. only one of the twenty- one aleph libraries studied kept the previous search terms in the search box on the results page. fourteen of the aleph libraries retained neither the previous search strategy nor the search terms. six libraries placed the search box at the bottom of the search results page, which could be easily missed. . post-search limit functions were not always readily available sometimes keyword searches produce an overwhelm- opac sorting options for keyword search relevance year (publication date) author title call # (subject) format default aleph year/author: title/year ascending: title: system sort: horizon publication date: title: author(ascending): system sort: millennium date: title: relevance: system sort: unicorn w descending ascending new to old: relevance: (ncsu) system sort: voyager kw (r) kw (b) descending ascending relevance: kw with relevance: system sort: figure . arl libraries sorting options for keyword search (as of march ) information technology and libraries | march information technology and libraries | march ing number of search results. since the relevance ranking functionality currently provided by ils vendors does not work very well, the best way to refine searches is to make effective search limit options available. limiting options such as format, language, date, availability, and location should be readily available on the results page. some ilss in our study hid this feature, either under a modified search link or an advanced search link. this made refin- ing a search unnecessarily cumbersome. . item statuses were not available on the search results page in addition to bibliographic information, users also need to know whether an item they want is available. having the item status on the browse page is very helpful because users can skip the records that have been checked out. some libraries studied did not have this informa- tion on the results browse page. users needed to go to the individual bibliographic records to find out whether an item was available or not. a few libraries provided an added-value option to limit the results by “available items”—a very useful feature. . a lack of value-added information a book cover image conveys an impression of a book that words cannot. it can also help a user recognize a book he or she has seen previously. in addition to cover images, libraries can provide value-added and contextual information by linking those images to tables of contents, summaries, sample passages of text, and reviews. one way libraries provide value-added and contextual infor- mation is to link cover images to the library of congress’s table of contents page. another way is to link opacs to information obtained from syndetics.com, a company that provides cover images, tables of contents, summa- ries, author biographical information, and reviews. the ohio state university library not only adds the table of contents into the marc record, but also links the names of the authors of a particular resource to other works by the same authors. this is a great discovery tool for find- ing related resources, and it is especially helpful, since in the future opacs will be able to search not only books but also articles and other resources. . title links were misleading we found that several libraries’ opacs title links on the results page did not take users to the detailed biblio- graphic record, but instead directed users to an alphabeti- cal title-browsing page. to get to the actual bibliographic record, users had to click a “display full record” link (which is sometimes difficult to locate) to view the indi- vidual bibliographic record. this misleading feature makes the retrieval process inefficient. . switching between individual records and the results list was cumbersome after viewing an individual bibliographic record, users will want to return to the results browse page, either by hitting the “back” button or by clicking on a “return to results” link. many library opacs in our study returned the user to the top of the results page rather than to the location to which the user had previously scrolled. this forced the user to scroll back down through the records that had already been examined. this feature ought to be improved. . the color of entry links that had already been read were not differentiated for over a decade now, web browsers have changed the color of links that have already been clicked on. however, this has not been the case with opacs. to solve this problem, visited bibliographic entry links on search results pages should likewise be given a different color from entries that have not yet been visited. this feature facilitates the browsing of the search results. if what has been viewed is clearly marked, users only need to focus on entries that have not yet been visited. some libraries in our study did not have this feature. . searched keywords were not highlighted when a keyword search is performed, highlighting the entered keywords in each bibliographic record that has been retrieved is helpful. based on the bibliographic elements in which the highlighted keywords appear, users can then decide how relevant the retrieved publica- tion is to their research. all five ils vendors provide this feature, and many libraries did a good job of implement- ing it. however, some libraries neglected to make this feature available. . many libraries lack a meaningful call-number browse feature library opacs should take better advantage of call number links by allowing users to browse them much as if they were browsing shelves in the stacks. to that end, opacs should link call numbers directly to a page with more useful identifying information, such as the authors and titles. no aleph library opacs that we studied cur- rently have this feature. instead, clicking on the hyper- linked call number field only leads users to a list of more call numbers, which is not helpful at all. . title link, subject link, and author link should be relabeled to be meaningful to end users (other value- added features) millennium’s “similar records” and voyager’s “more like this” are added to pull similar titles under the same article title | author revitalizing the library opac | mi and weng subjects. unicorn and horizon offer a panel on the left side of the detailed book record, which can add meaning- ful information to these links. but how the panel is used depends on the individual libraries. some libraries use the panel with only library holding information, but other libraries, such as university of virginia, make an infor- mative presentation of those links to students. virginia has added three browse features to make the index links much more meaningful: “find more by this author” (author link), “find more on these topics” (subject link), and “nearby items on shelf” (call number link). (see figure .) this value-added feature can indeed facilitate retrieval process. by analyzing five major integrated library systems’ opacs among arl libraries, the authors have come to believe that librarians can make a big difference in improving opacs. no matter how good the library system is, librarians still need to invest effort, time, and technical knowledge to configure and take full advantage of the many capabilities that ilss offer. public services, technical services, and system librarians should all work together to continuously study the usability of opacs and to make them more effective. it is true that all current opacs lack spell-check and automatic stemming func- tionality. aleph and horizon need to add relevance rank- ing, and millennium, unicorn, and voyager should make our data work harder and relevance ranking algorithms more effective. besides those systems in need of improve- ments, the study shows that all library opacs could do a much better job if they focus on the user’s needs. ■ the opac bibliographic display study when the web opac was introduced, libraries around the world quickly abandoned the traditional card cata- log display and adopted the line-by-line display with display labels on one side and bibliographic information on the other. because the line-by-line display format can be locally customized, each library’s opac bibliographic display looks very different. for decades, most academic libraries in the united states have used aacr and marc as their content and metadata standards for resource description and access. marc and aacr were originally created for card catalogs in which descriptive elements and access elements were separately defined and presented. the line between the two types of elements has become less distinct in today’s web environment. many elements in bibliographic records can serve as both description and tracing elements on opacs. hyperlink functionality has also streamlined the retrieval process. to see how academic libraries in the united states format their opac bibliographic displays, the authors examined the opacs of fifteen academic libraries. the purpose was to study the effectiveness of the display of records in different formats. in the mid- s, wool stud- ied the bibliographic display practices for monographs of thirty-six online catalogs in the united states. in his study, five criteria were used to analyze each bibliographic record structure. the authors of this paper adopted for analysis three of the five opac bibliographic display criteria used by wool, only this time with an emphasis on the user’s perspective and needs. eight different titles were reviewed and compared: three monographs, two serials, one video recording, and two sound recordings. the analysis given below is based on the following three criteria: ■ the accuracy and clarity of display labels; ■ the order of bibliographic elements display; and ■ the utilization of bibliographic data. accuracy and clarity of display labels for this discussion, the authors divided the bibliographic elements into three areas: ■ the first tier: information about author/contributor, title, imprint, and subjects; ■ the second tier: other descriptive information, including the physical description, notes, related con- tributors, related titles, etc.; and ■ the third tier: the linking fields (marc x– x fields) and the electronic location and access field (i.e., field). the first-tier elements the information displayed in the first tier can be consid-figure . university of virginia libraries catalog information technology and libraries | march information technology and libraries | march most libraries in our study used the label “author” for the principal author. the principal author could be a personal author, a corporate author, or a conference name. if it is a personal author, it could be a writer, an artist, or a composer. some opacs used “author” to represent all types of responsible bod- ies, be it a personal author, a corporate author, a meeting name, an artist, a music com- poser, etc. this use of a single label to cover a diverse set of situations is confusing. some libraries used separate labels (“author,” “corporate author,” “meeting name,” “author/ artist,” “author/composer,” or “author, etc.”) for different types of responsible bodies (see appendix d). “uniform title” was defined in aacr to collocate resources derived from the same original intellectual or artistic creation. for example, when cataloging a transla- tion, in addition to its official translated title, an estab- lished uniform title is entered to indicate the original work. when browsing by uniform title on a properly set opac, all entries related to the original intellectual creation should be retrieved. this uniform title brows- ing feature helps users locate related publications in the catalog. the problem is that the term “uniform title” is only understood by catalogers, not by others. there is no label for such an entry that can be easily understood by the average user. however, suppressing the uniform title entry to avoid confusing users will cause the opac to lose its helpful collocation functionality. some libraries studied use the term “uniform title” as a display label. some libraries use “other title” as a display label. some libraries display this entry under the label “title” along with the title proper (title in the field). none of the above-mentioned arrangements are ideal. the display labels for subject headings provided by each library were very similar. most academic libraries in the united states use the library of congress subject headings and the medical subject headings as the thesauri for subject entries. specifying the thesauri for headings on opacs with acronyms like “lcsh” and “mesh” is of no help to users, because these thesauri do not clarify anything that would assist users in their research. figure lists the display labels used by libraries in the study. ered the key elements for identification. opac users first examine them and decide if the manifestation described is relevant to their query. most opacs studied used “title” as the display label for the title statement. this element actually consists of the title and statement of responsibil- ity (author, etc. statement). using the label “title” alone is not inclusive enough. one library (university of arizona library) displayed only the title portion under the label “title” and provided a separate label, “author/con- tributor info,” for the statement of responsibility portion, which, while helpful in a limited way, could also create more confusion. let us consider, for example, the project directory (répertoire des projets) of tdc (in french, cdt). the title statement for this data would be “project direc- tory / tdc = répertoire des projets / cdt.” here, the english title and statement of responsibility is equiva- lently presented with its french title and statement of responsibility. the opac display using the university of arizona library’s model is as follows: title: project directory author/contributor info: tdc = répertoire des projets / cdt. this arrangement will not work for items with titles and statements of responsibility in multiple languages presented on a single manifestation. the french title appears under the label “author/contributor info,” which makes no sense. marc fields library of congress subject (marc field nd indicator ) medical subject (marc field nd indicator ) d is p la y l a b e ls subject (lcsh) subject (mesh) subject-lib. cong. subject-medical subject lc subject medical library of congress subject headings medical subject headings subject(s) subject(s) subject, general subject, geographic subject, medical subject med. subject figure . display labels for subjects article title | author revitalizing the library opac | mi and weng the second-tier elements the elements in the second tier include the physical description, notes, related authors, and related titles. this is an area where mapping bibliographic elements onto proper display labels is difficult. this area was also not managed well by the libraries studied. unlike first-tier elements in which one element usu- ally corresponds to a unique display label, second-tier elements exhibit two patterns in the opacs exam- ined: many-to-one and one-to-many. that is, multiple categories of data (of different marc fields) can be represented by one display label, e.g., incorporating physical description, numbering notes, and publication numbering into “description” (many-to-one). on the other hand, one display label can represent one single, repeatable bibliographic element (the same marc field repeated many times), e.g., multiple general notes (one- to-many). both arrangements (one-to-many and many- to-one) can result in a simpler, cleaner public display, since some descriptive elements are self-explanatory and users can get by without specific display labels supplied. the disadvantage of these arrangements is that the level of specificity of public displays is com- promised. some important descriptions can be easily missed if they are clustered in a group of elements. for bibliographic elements that are not self-explanatory, this type of arrangement can fail to convey useful informa- tion, or even worse, deliver inaccurate or vague infor- mation. for example: description: v. : ill. ; cm (physical description, marc field) report year ends mar. . (numbering note, marc field) ’ – (publication span, marc field) published: philadelphia : robert morris associates, – (imprint, marc field) ’ – (publication span, marc field) annual (frequency, marc field) the numbering field (field ) is defined to describe a serial publication’s chronological or numerical publi- cation extent. carelessly placing data like “’ -” under labels such as “description” or “published” is very unclear. in fact, it is inaccurate because “’ -” is the pub- lication span, not the publication date. without a proper label, it is difficult to convey this information to users. some libraries we studied used such labels as “publica- tion history,” “publishing history,” “publication dates,” or “volume/date range” to describe the publication span. this practice is misleading (see appendix e). names like coauthors, editors, cast members, perform- ers, related corporate names, or meeting names of people who contributed to or were involved in the creation of the work are considered secondary contributors. using one label to cover the various roles (author, editor, composer, etc.) is the practice most libraries have adopted. like the primary author field, this element represents a variety of roles depending upon the type of manifestation. some opacs used one display label to cover all related per- sonal names, corporate names, and meeting names (see appendix f). most libraries failed to supply a proper label for a secondary name when it was entered with a related title. this so-called “name-title added entry” is provided to collocate materials under the same author and title in the catalog. ideally, the name-title combined element, pro- vided with redirect functionality via hyperlink, should perform an author-title combination search for exact retrieval. most opac systems could only perform either an author or a title search. the search results were unsur- prisingly irrelevant, because they did not utilize both elements of the name-title added entry to produce results that were sufficiently specific: users would get only a list of authors or a list of titles instead of an author-title com- bination entry list. some libraries presented this type of element only as an unhyperlinkable note, which defeats the purpose of having such data available. handling series for opac displays is also chal- lenging. the majority of opacs studied did a poor job in this area. in general, a series title transcribed from the resource also functions as an access element if the transcribed title is the same as the established one in the authority file. when the transcribed series title is different from the established series title, ideally the transcribed series title should only be accessible via the library system’s cross-references feature, which then directs users to bibliographic records that contain the established entry. this type of descriptive element is not meant to be displayed on the opac. the opacs examined used the labels listed in figure to handle transcribed and established series entries. labels listed in the same row were taken from the same opac. as can be seen, users are not expected to know the difference between a “series statement” and a “series.” in many cases, these two elements are identical due to the vendor authority control process. this could confuse the user, especially when both elements are displayed right next to each other. information technology and libraries | march information technology and libraries | march the third-tier elements the third-tier elements consist of linking fields (marc xx fields) and electronic location and access fields (marc field). the linking fields are used mostly in serial bibliographic records. their purpose is to link the title being described to its related publications, e.g., supplements, translations, preceding titles, or succeeding titles. elements in this category should be displayed and linked directly to the related record via control numbers provided in the bibliographic record. if the catalog does not have the related record, a clear message should indi- cate this to the user. unfortunately, many libraries do not display all the linking entries. none of the opacs stud- ied offered direct link functionality. instead, what was usually offered was a redirect feature via hyperlink that prompted the system to issue a new author or title search. the direct link functionality via record control numbers was never made available. if the library did not have the related entry, the opac system simply took the user back to the original entry—a very confusing design flaw. to ease the user’s access to internet resources, the electronic location and access element (marc field) was defined for catalogers to record the internet location of the resource being described and its related informa- tion. by clicking the hyperlinked element on an opac, users seamlessly get to the desired electronic document site. the url specified in the field might link to full-text documents, the table of contents, the document abstract, the publisher’s description, or the author’s biographi- cal information. a label that fits all types of materials is crucial. the bibliographic elements displayed under the label should also be carefully managed. under the label, some libraries displayed the type of resource (e.g., table of contents). other libraries displayed the http url only. some libraries displayed both the type of resource and the http url (see figure ). as for the location of the label in the opac record, we found that the loca- tion of the url link depended on the opac in which it appeared: in some opacs, links were located at the top of records; in others, they appeared in the middle or at the bottom. we found that the location of the link was not ter- ribly critical, provided that the label was prominent and the display text understandable. the order of the bibliographic elements display the way bibliographic data is organized in each opac record, together with display labels, helps users to quickly identify library resources. although each library can locally choose the arrangement of bibliographic data displayed on its opac, most libraries prefer to place the citation information (author, title, publication) ahead of other elements. the sequence of the other elements exhibited enormous variation in the opacs studied. some libraries placed the electronic access element above all other data (suny buffalo); some libraries placed local holdings information, call number, and item availability in the middle of the bibliographic record. arrangements were clearer and more understandable when provided with clear labels and a distinct layout between the local holdings information and bibliographic data. problems arose when second-tier elements were mingled with first- tier elements and when they shared the same display label. see example in figure . in this example, two titles are displayed under the “title” label. the first title, “rma annual statement stud- ies,” is the full title (marc field ) of the publication. the second title, “rma annual statement studies: industry default probabilities and cash flow measures,” is the title of the resource’s related publication (marc field ), which normally is considered a second-tier element and should be placed farther from the title proper with a clear label. since the display order of bibliographic elements is completely customizable, we found in our study that few libraries put enough effort into providing clear bibliographic displays. more importantly, records in different formats (e.g., mono- graphs, serials, music materials, video recordings) were not given equal attention. some labels and data sequences might work for one format, but not another. utilization of bibliographic data another factor that has an effect on the usability of an opac is the utilization of bibliographic data. two issues are addressed in terms of utilization of bib- liographic data: ( ) the completeness and suitability of the metadata displayed on an opac, and ( ) the extent of repurposing the bibliographic data and creating added value to an opac. a typical bibliographic record contains descriptive data, access data, and admin- label for transcribed element label for established element series statement series series statement series indexed as other series series series note series description series figure . display labels for series article title | author revitalizing the library opac | mi and weng istrative data. descriptive data is provided to describe the manifestation cataloged and is considered of inter- est to the public. access data is entered and indexed for retrieval. administrative data is used for setting up search limits (e.g., limit by language, format) and pull- ing statistics (e.g., how many titles in spanish). it is most useful for internal, administrative use. librarians must be careful when deciding whether such data elements will be displayed. in terms of the completeness and the suitability of metadata in the opac display, the authors discovered the following in the opacs studied: . many libraries’ opacs displayed control numbers, such as the oclc control number (the field), the lc control number ( field), and other local system control numbers. this type of information is usually of no interest to the public. see example in figure . in this example, the numbers listed under the label “wln #” represent different types of system control numbers, which are of no concern to users and therefore should not be displayed. . some opacs displayed bibliographic data from the leader fields of the cataloging record. marc leader fields are a group of fixed-length codes that represent the type of resource (monograph, serial, or musical score) and material format (print, elec- tronic, or sound recording). the information could be helpful for patrons if they are displayed with the proper label on the opac. libraries that chose to display the leader data on their opacs did not do a good job of making the information clear to users. for example, one library listed “journals and newspapers,” “computer file,” “serial,” “book,” “e-resource,” and “gov publication” under the label “record type” (see figure ). seeing so many record types under one label can easily confuse library users. . some libraries omitted certain crucial variable fields, e.g., the linking entry complexity note (field , containing information about title history), related title access entries (fields and , containing related titles), and linking entries (link- ing the record to other bibliographically related records, e.g., x, x, and x fields). these fields are defined with a clear purpose and should be carefully considered for public display with clear labels. some libraries in our study displayed them but left other irrelevant information on the opac, which clutters the display with information that does not help users. see example in figure . in this example, under the label “related publica- tion,” the french version and the spanish version of jama are displayed. in addition to the french title and the spanish title, the marc language code and its corresponding issn are also displayed. the language code and the eight-digit issn number— since no separate label is provided for them—are confusing. . the linking elements not only should be displayed on the opac, but should also be hyperlinkable. they ought to be used to link to related biblio- graphic records. in an online environment, this sort of field can also function as a descriptive element. some opacs displayed linking entries but did not enable hyperlink functionality. some libraries dis- played two instances of them, one as a descriptive element and the other as a linking element with hyperlink capability. another important aspect of making use of bib- liographic data is repurposing the bibliographic data to provide added value to opacs. lorcan demsey mentions frequently in his blog that in order to sustain library value, libraries should “make data work harder.” he points out that “libraries have invested a great deal in bibliographic data—yet it has remained somewhat inert in our catalogs, failing to release the value of the investment.” these rich data can be better utilized for different purposes, including designing an enhanced opac. lavoie, et al. described further in their recent article about data mining: as more activities move into networked spaces, more areas of our lives are shedding data. this data is increasingly being mined for intelligence that drives services. . . . [c]ompanies like amazon repurpose data to create added value. this is a lesson librarians must learn if they want to improve their own visibility and value in increasingly crowded digital information spaces where users, as always, want good results with- out too much time or effort. . . . the good news is that libraries don’t come to this task empty-handed but with figure . online opac record from suny buffalo figure . online opac record from the college of new jersey information technology and libraries | march information technology and libraries | march rich, structured information about the materials in our collections. tim o’reilly highlighted in his article the successful example of how amazon reutilizes data: amazon relentlessly enhanced the data, adding pub- lisher-supplied data such as cover images, table of contents, index, and sample materials. even more importantly, they harnessed their users to annotate the data, such that after ten years, amazon, not bowker, is the primary source for bibliographic data on books, a reference source for scholars and librarians as well as consumers. . . . effectively, amazon “embraced and extended” their data suppliers. all opacs reviewed in the study operate within the traditional vendor-supplied module. this long-estab- lished approach gives libraries limited flexibility to cus- tomize the search key options, search results displays, restricted sorting options, and pre- and post-search limit options of their opacs. unfortunately, libraries can do very limited data mining inside the vendor’s hard-coded framework. many valuable metadata are buried in the bibliographic database. system vendors have failed to make the most of technology to better utilize data. very few libraries have thought outside the box and taken advantage of the existing rich bibliographic data. the emergence of north carolina state university’s endeca- powered opac was a good example of repurposing data and creating value-added information. the data facets used on ncsu’s single search-and- browse combined opac interface are pulled and repur- posed from their sirsi/dynix database. as one might have expected, eight of the eleven facets are extracted from the library’s marc bibliographic records (“availability” and “browse: new” are from item records). out of the eight fac- ets, four are extracted from subject headings; two are from the fixed fields; one is from the call number field and one from the variable fields of the bibliographic record. ■ discussion and recommendation based on the authors’ findings above, the following are the primary factors that have contributed to the ineffectiveness of the opacs offered by today’s academic libraries. . system limitations the inadequacy of today’s ils has been a known problem. inflexible search options make library cata- logs difficult to use. despite the fact that some vendors diligently enhance their systems’ functionalities, overall performance is still disappointing. karen markey pointed out in a recent article that one of the reasons why the solu- tions recommended by researchers in the s were not applied to online library catalogs was “the failure of ils vendors to monitor shifts in information-retrieval tech- nology and respond accordingly with system improve- ments.” antelman et al. observed similarly that all major ils vendors are still marketing catalogs that represent second-generation functionality. despite between-record linking made possible by migrating catalogs to web interfaces, the underlying indexes and exact-match boolean search remain unchanged. it can no longer be said that more sophisticated approaches to searching are too expensive computationally; they may, however, to be too expensive to introduce into legacy systems from a business perspective. since ils vendors first introduced their products back in the s, user behavior and expectations have changed immensely. while libraries have started to figure . online opac record from the university of washington. figure . online opac record from the university of michigan.figure . online record from suny buffalo article title | author revitalizing the library opac | mi and weng recognize the changes and are working hard toward meeting the needs of multiple generations of users, little can be done if ils products still operate within the same old-fashioned information-retrieval structure. because ils vendors have failed to revamp their opac modules to meet user needs, libraries have been forced to seek other options. north carolina state university is one of the first libraries to exercise its options. its new opac system, powered by endeca (operated on the sirsi/dynix platform), has shown remarkable improvements in ease of use, which usability tests have verified. recently, two ils vendors (innovative and ex libris) have been in the process of developing new opac modules using new technology and a new approach in data mining. . libraries are not fully exploiting the functionality already made available by ilss unsurprisingly, the opacs examined by the authors, if powered by the same vendor, showed similarities in general layout and interface features. during the study, it soon turned out to be easy for the authors to recognize the ils system of each opac. as mentioned previously, we expected opacs to vary somewhat. what was unexpected was the huge differences in, among other things, interface layout, search options and search languages, behind-the- scenes search algorithms, search results displays, display labels and the corresponding bibliographic data, and what data was chosen for display. the disparities that we found in these features suggested that there had been great differ- ences in the amount of attention, energy, and time devoted by each library to designing its opac. some libraries took advantage of available features and made better use of them than others. (see appendix g for examples of best practices of library opacs.) many libraries did only the very minimum. while we recognize that academic library opacs are difficult to use, we also need to recognize that some libraries do not fully exploit existing resources, thereby exacerbating the difficulty of using their opacs. . the unsuitability of marc standards to online bib- liographic display as previously mentioned, aacr and marc were initially designed for card catalogs without display labels in mind. many marc fields can be used for multiple purposes. providing labels that properly fit all the cata- loging data needed to cover all types of resources is nearly impossible. from the opacs studied, some librar- ies used vague labels in an effort to encompass as many circumstances as possible. some libraries used labels suit- able only for certain formats, but not all formats. neither approach is satisfactory. the solution has to come from cataloging and metadata standards. wool identified this issue back in the s: the interchangeability of descriptive data elements and access points (since each can be made to serve both functions online) makes the separate creation of description and headings seem pointless and bur- densome. labeling of data elements (made possible through the mapping of terms to marc fields) creates a need for simpler, less ambiguous bibliographic data definitions than are appropriate for the dense and context-rich narrative-style records catalogers continue to create . . . cataloging standards will need to be rewritten in order to provide the kind of data flexibility expected in online systems . . . records flexible enough to be added to, subtracted from, and rearranged without loss and gar- bling of meaning. what is needed is a modular record structure, in which every segment of data can stand on its own with appropriate labeling and which can sup- port all possible display lengths and combinations of data elements. a decade later, not much progress has been made in improving cataloging and metadata standards for online display. while enhancing cataloging and metadata stan- dards for better retrieval is desirable, making the stan- dards more complicated and difficult to adopt in order to accommodate opac displays is not. as librarians are working to simplify cataloging, our essential rich metadata should not be sacrificed. one possible solution is to have the system recognize the existence of certain subfields and produce specific display labels accordingly. this certainly will not solve all the issues with regard to display labels. regardless, there is much room for improvement, and librarians’ attention is this area is critically needed. ■ conclusion the information-seeking world has entered an era of self- service. roy tennant described well the self-service trend: “i wish i had known that the solution for needing to teach our users how to search our catalog was to create a system that didn’t need to be taught.” tim o’reilly also indicated in his article “what is web . ” that “the web . lesson [is to] leverage customer-self service and algorithmic data management to reach out to the entire web, to the edges and not just the center, to the long tail and not just the head.” he also argued that “[t]rusting users as co-devel- opers” is one of the core competencies of web . compa- nies. academic libraries should aim toward designing a user-centered, self-sufficient, twenty-first-century online catalog that fits the web . model. the ultimate goal is that users will be comfortable and confident using library opacs for their information needs wherever a computer information technology and libraries | march information technology and libraries | march is available and without special training. as campbell and fast have trenchantly asked, “are we witnessing a major disruption, a large-scale redefi- nition of information design and delivery so radically different from the traditional library environment that it renders irrelevant all our experience in bibliographic con- trol?” this remains an open question. regardless, a new generation of opacs will need to be in place soon. much needs to be done to make academic library opacs mat- ter. academic librarians cannot afford to be considered irrelevant in the information-seeking world. the future of academic libraries relies on effective opacs. this is one of the most pressing tasks that must be accomplished. references and notes . cathy de rosa et al., perceptions of libraries and information resources: a report to the oclc membership (dublin, ohio: oclc, ), – . http://www.oclc.org/reports/ perceptions.htm (accessed jan. , ). . karen coyle and diane hillmann, “resource description and access (rda): cataloging rules of the th century,” d-lib magazine , no. / ( ). http://www.dlib.org/dlib/janu- ary /coyle/ coyle.html (accessed feb. , ). . anna m.van scoyoc and caroline cason, “the electronic academic library: undergraduate research behavior in a library without books,” portal: libraries and the academy , no. ( ): – . . carol tenopir, “user and users of electronic library resources: an overview and analysis of recent research stud- ies,” council on libraries and information resources, . http://www.clir.og/pubs/reports/pub /pub (accessed jan. , ). . cathy de rosa et al., the oclc environmental scan (dublin, ohio: oclc, ), http://www.oclc.org/reports/ escan/introduction/default.htm (accessed jan. , ). . d. grant campbell and karl v. fast, “panizzi, lubetzky, and google: how the modern web environment is reinventing the theory of cataloguing,” the canadian journal of information and library science , no. ( ): – . . roy tennant, “breaking library services out of the box,” presentation ( ), http://www.cdlib.org/inside/news/ presentations/rtennant/ netspeed/ (accessed feb. , ); andrew pace, “my kingdom for an opac,” american libraries online (feb. ), http://www.ala.org/ala/alonline/ techspeaking/ colunms/techfeb .cfm (accessed feb. , ); karen g. schneider, “how opacs suck, part : relevance rank (or the lack of it),” ala techsource blog (mar. , ), http://www.techsource.ala.org/blog/ / /how-opacs- suck-part- -relevance-rank-or-the-lack-of-it.html (accessed feb. , ); karen g. schneider, “how opacs suck, part : the checklist of shame,” ala techsource blog (apr. , ), http:// www.techsource.ala.org/blog/ / /how-opacs-suck-part- -the-checklist-of-shame.html (accessed feb. , ); “how opacs suck, part : the big picture,” ala techsource blog (may , ), http://www.techsource.ala.org/blog/ / / how-opacs-suck-part- -the-big-picture.html (accessed feb. , ); lorcan dempsey, lorcan dempsey’s weblog (oct. , ), http://orweblog.oclc.org/archives/ .html (accessed feb. , ); kristin antelman, emily lynema, and andrew k. pace, “toward a twenty-first century library catalog,” infor- mation technology and libraries , no. ( ): – . . roy tennant, “libraries through the looking-glass,” ala midwinter endeavor presentation. http://www.cdlib. org/inside/news/presentations/rtennant/ ala/ (accessed march , ). . charles ammi cutter, rules for a printed dictionary cata- logue (washington, d.c.: government printing office, ). . d. grant campbell and karl v. fast, “panizzi, lubetzky, and google: how the modern web environment is reinventing the theory of cataloguing,” . . holly yu and margo young, “the impact of web search engines on subject searching in opac,” information technology and libraries , no. ( ): . . jakob nielsen, “mental models for search are getting firmer,” in jakob nielsen’s alertbox, http://www.useit.com/ alertbox/ .html (accessed feb , ). . eng pwey lau and dion hoe-lian goh, “in search of query patterns: a case study of a university opac,” informa- tion processing and management , no. ( ): – . . holly yu and margo young, “the impact of web search engines on subject searching in opac,” . . dinet jérome, favart monik and passerault jean-michel, “searching for information in an online public access cata- logue opac: the impacts of information search expertise on the use of boolean operators,” journal of computer assisted learning , no. ( ): – . . gregory wool, “the many faces of a catalog record: a snapshot of bibliographic display practices for monographs in online catalogs,” information technology and libraries , no. ( ): . . the fifteen libraries are located at the college of new jersey, library of congress, northwestern university, princeton university, state university of new york at buffalo, temple uni- versity, university of arizona, university of florida, university of illinois–urbana-champaign, university of michigan, univer- sity of minnesota, university of rochester, university of texas– austin, university of washington, and vanderbilt university. . gregory wool, “the many faces of a catalog record: a snapshot of bibliographic display practices for monographs in online catalogs,” – . . eight titles representing monograph, serial, video record- ing, and sound recording were used to study the effectiveness of the bibliographic display. the eight titles are: ( ) to love the wind and the rain: african americans and environ- mental history, edited by dianne d. glave and mark stoll. university of pittsburgh press, . (monograph) ( ) to kill a mocking bird, by harper lee (mongraph) ( ) rma annual statement studies, robert morris associates, - (serial) ( ) sideways ( th century fox, ) (video recording) ( ) chamber music (newport classic, ) (sound recording) ( ) end of summer book of hours ; bright music, naxos, / by ned rorem (sound recording) ( ) jama : the journal of the american medical association, - (serial) article title | author revitalizing the library opac | mi and weng ( ) the st century at work, by lynn a. karoly (rand, ) (mongraph) . many vendors retag the field to in bibliographic record and create an field based on the contents of the field. the series title in the field receives authority control. many libraries prefer not to restore the field back to the fields causing the duplicate series statements on opac if both fields are displayed. . lorcan demsey, “making data work—web . and cata- logs.” . ibid. . brian lavoie, lorcan dempsey, and lynn silipigni con- naway, “making data work harder,” library journal.com (jan. , ), http://www.libraryjournal.com/article/ca . html (accessed jan. , ). . tim o’reilly, “what is web . : design patterns and business models for the next generation of software,” (sept. , ), http://www.oreillynet.com/pub/a/oreilly/tim/ news/ / / /what-is-web- .html (accessed jan. , ). . tito sierra, “a faceted interface to the library cata- log,” ala midwinter meeting, http://www.lib.ncsu.edu/ endeca/presentations.html (accessed feb. , ). . karen markey, “the online library catalog: paradise lost and paradise regained?” d-lib magazine , no. / ( ). http://www.dlib.org/dlib/january /markey/ markey.html (accessed feb. , ). . kristin antelman, emily lynema, and andrew k pace, “toward a twenty-first century library catalog,” . . gregory wool, “the many faces of a catalog record: a snapshot of bibliographic display practices for monographs in online catalogs,” – . . roy tennant, “lipstick on a pig,” library journal.com (apr. , ), http://libraryjournal.com/article/ca . html (accessed feb. , ). . tim o’reilly, “what is web . : design patterns and busi- ness models for the next generation of software.” . d. grant campbell and karl v. fast, “panizzi, lubetzky, and google: how the modern web environment is reinventing the theory of cataloguing,” . appendix a. default search keys used by arl libraries (as of march ) appendix b. keyword search keys used by voyager libraries keyword (relevance) keyword (boolean) keyword with relevance ranking keyword (enclose phrases “in quotes”) keyword anywhere (user “” for phrase) keyword combined (use and/or/not “ “ for phrase) keyword anywhere (relevance ranked) keyword (and or not) keyword anywhere advanced boolean words anywhere keyword boolean basic keyword keyword(s) (user and, or, not, or “a phrase”) any word anywhere boolean search (use and or not) relevance keyword (user + for key terms) command keyword keyword phrase keyword (use “and” “or” “not”) keyword and or not( keyword boolean) keyword (results sorted by relevance) expert keyword keyword keyword expert (user an or not “phrase”) keyword command ranked keyword keyword information technology and libraries | march information technology and libraries | march keyword (ranked by relevance) keyword keyword command search find all words search for a phrase keyword (quick search) boolean search appendix c. default keyword search help page provided by voyager system keyword search ■ enter words and/or phrases ■ use quotes to search phrases: "world wide web" ■ use + to mark essential terms: +explorer ■ use * to mark important terms: *internet ■ use ? to truncate (cut off) words: theat? finds theaters, theatre, theatrical, etc. ■ do not use boolean operators (and, or, not) to combine search terms boolean ■ use the boolean terms (and, or, not) to combine search terms. ■ use quotation marks to search for a phrase, e.g., "united states" ■ use ? to truncate a word, e.g., browser? ■ use parentheses to group search terms, e.g., (automobile or car) and repair appendix d. display labels for entries of principal responsibility marc fields libraries (personal name) (corporate name) (meeting name) u. of arizona author author author u. of ill. author author conference lc personal name corporate name meeting name u. of minnesota author author author u. of michigan author author author northwestern u. author, etc. author, etc. author, etc. princeton u. author/artist author/artist author/artist u. of washington author author author suny buffalo author author author temple author corp author conference u. of florida author, etc. author, etc. author, etc. u. of rochester main author main author conference ut austin author corporate author conference tcnj principal author principal author conference name vanderbilt u. author corporate author meeting/event name article title | author revitalizing the library opac | mi and weng appendix e. display labels for publication extent libraries marc field u. of arizona issued u. of ill. publication history lc description u. of minnesota published u. of michigan pub history northwestern u. extent of publication princeton u. description u. of washington (suppressed from opac) suny buffalo publication dates temple publication started u. of florida publishing history u. of rochester (suppressed from opac) ut austin publication coverage date tcnj description vanderbilt u. volume/date range appendix f. display labels for entries of secondary responsibility marc fields libraries (personal name) (corporate name) (meeting name) u. of arizona other auth other auth other auth u. of ill champaign other name other name other name lc related names related names related names u. of minnesota contributor contributor contributor u. of michigan contributors - people contributors - other contributors - other northwestern u. other authors, title, etc. other authors, title, etc. other authors, title, etc. princeton u. related name(s) related name(s) related name(s) u. of washington alt author alt author alt author suny buffalo contributors contributors contributors temple other author(s) other author(s) other name u. of florida other author(s), etc. other author(s), etc. other author(s), etc. u. of rochester other author(s) other author(s) other author(s) ut austin added author (not display) (not display) tcnj other contributor(s) other contributor(s) conference name vanderbilt u. author, editor, etc. corporate author meeting/event information technology and libraries | march information technology and libraries | march appendix g. examples of best practices of opacs (accessed july , ) search interface, including retaining search keys and searched terms university of notre dame http://alephprod.library.nd.edu: /f/?func= find-b- &local_base=ndu pub keyword searching ability michigan state university http://magic.msu.edu/search~/x facets browsing (endeca) north carolina state university http://www.lib.ncsu.edu/catalog mcmaster university http://libcat.mcmaster.ca make author, subject and call number links more accessible university of virginia https://virgo.lib.virginia.edu/uhtbin/cgisirsi/ / uva-lib/ / / /x links to amazon ratings ohio state university http://library.ohio-state.edu/search direct export to refworks johns hopkins university https://catalog.library.jhu.edu/ipac /ipac. jsp?profile=default#focus university of chicago http://libcat.uchicago.edu/ipac /ipac. jsp?profile=ucpublic cover art/toc/ summary/review indiana university http://www.iucat.iu.edu/authenticate.cgi?status=start guesstimate/del.icio.us persistent link enabled virginia tech http://addison.vt.edu information technology and libraries | march information technology and libraries | march touchable online braille generator wooseob jeong a prototype of a touchable online braille generator has been developed for the visu- ally impaired or blind using force feedback technology, which has been used in video games for years. without expensive devices, this prototype allows blind people to access information on the web by touching out- put braille displays with a force feedback mouse. the data collected from user studies conducted with blind participants has pro- vided valuable information about the opti- mal conditions for the use of the prototype. the end product of this research will enable visually impaired people to enjoy informa- tion on the web more freely. the united states has made some attempts to nationally address infor- mation access for those with disabili- ties. section of the rehabilitation act (www.section .gov) requires federal agencies to make their elec- tronic information accessible to peo- ple with disabilities, mainly those who are visually impaired. the library of congress launched a web- braille service (www.loc.gov/nls/) for the blind in , which continues today. with the upsurge in infor- mation stored on the internet, the importance of these issues cannot be overemphasized. many products have been devel- oped to help the visually impaired use technology. several braille out- put and input devices are available, such as the braille notetaker (www. artictech.com) and voice synthesizers for screen readers like jaws (www. freedomscientific.com/fs_products/ software_jaws.asp). while these products are mainly for textual information, recent devel- opments put more focus on graphi- cal displays. the american national institute of standards and technology proposed a “pins” down imaging system for the blind (www.nist.gov/ public_affairs/factsheet/visualdis- play.htm). uniplan in japan and ksg america (www.kgs-america.com/ dvs.htm) have produced other prod- ucts based on similar ideas. software like the duxbury braille translator (www.duxburysystems.com) can translate plain text into braille out- put, which can then be used for embossed printing. however, such products are fairly expensive, rang- ing from hundreds to several thou- sands of dollars in addition to the cost of computers. fortunately, there is a potentially promising solution. based on the technology used in prior research, it is possible to develop an online braille generator. the braille could then be read either by touching the screen with a fingertip sensor or through the use of a force feedback mouse similar to the type used in some video games. this application has several advantages over existing devices. first, it does not require expensive special devices—only a $ mouse, which is readily available. also, the technology is available as long as there is access to the internet. another advantage is that this technology utilizes the existing braille skills of visually impaired people. the same technology can be used for produc- ing image displays as well, allowing for the creation of a virtual museum for the blind where they can touch objects that are displayed alongside their braille descriptions. literature review force feedback has been studied under the name of haptic perception. haptic perception involves sensing the movement and position of joints, limbs, and fingers through kinesthe- sia and proprioception, and sensing information through the skin’s tactil- ity. haptic output can be achieved through several techniques, including pneumatic, vibrotactile, electrotactile, and electromechanical stimulation. this study examines only vibrotac- tile haptic output methods because vibrotactile stimulation is easily cre- ated, manipulated, and delivered. it is also easily perceived by users through the use of commonly avail- able software and devices. researchers have begun to develop various haptic input/output devices and software, such as massachusetts institute of technology’s (mit) fre- quently used phantom haptic inter- face. along with these developments, a number of studies have tried to apply haptic displays to real-world computing, including a force feed- back braille system, force feedback virtual reality modeling language (vrml), a force feedback x window system , and gis. haptic studies have only recently become more mainstream, and there are few extensive studies with real subjects. gillespie and others devel- oped the “virtual teacher,” a device for manual skill learning, which they tested with participants and found that most profited from the “force feedback teacher.” langrana and others used the rutgers master ii, a dexterous, portable master for virtual reality simulations for force feedback using four fingers. in their experiment of tumor detection in virtual livers with subjects, the experimental group with force feedback training performed slightly better than the control group. this may mean that either the training methods need improvement or that the task did not require extensive training. colwell and others confirmed that a hap- tic interface (impulse engine ) has considerable potential for blind computer users through their three- dimensional objects experiment with subjects. jeong tested ordering communications wooseob jeong (wjj @uwm.edu) is associate professor at the school of information studies, university of wisconsin–milwaukee. article title | author touchable online braille generator | jeong tasks in auditory and haptic dis- plays with subjects and found that subjects performed better with haptic-only displays than with audi- tory-only displays or with auditory/ haptic combination displays. several studies already attempted to apply force feedback technology to assist blind people’s computing. ramstein conducted a pilot study to apply haptics to braille. yu and brewster compared the use of force feedback in multimodal virtual real- ity and printed medium in visu- alization for the blind. tzovaras and others tried to implement a virtual reality interface with force feedback for blind people. ramloll and others studied the use of haptic line graphs with sound for blind students. emery and others tested a multi- modal haptic interface with older adults to find that all participants per- formed well under auditory-haptic bimodal feedback. jacko and others tested a multimodal interface with normal vision older adults and visually impaired older adults, finding that in some cases, nonvisual feedback forms—including auditory or haptic feedback—demonstrated significant performance gains over the visual feedback form. s. jeong and others proposed an interactive system that combines an immersive virtual environment with a human- scale haptic interface. when conducting user studies with the visually impaired, it is nec- essary to separate the completely blind from the partially sighted. in spite of the different characteristics of these two groups, the literature on visually impaired people typi- cally does not distinguish between them. this distinction is especially important if the legally blind or those with low vision are included in the definition of visually impaired. the challenges to the partially sighted are different from those of the totally blind, demanding different assistance and considerations. in fact, the completely blind rep- resent a small portion of the visually impaired population. according to an advisor in wisconsin’s division of vocational rehabilitation, less than percent of her advisees are totally blind and require very specialized attention quite different from par- tially sighted people. purpose of study the purpose of this study is to explore the feasibility of using force feedback technology to facilitate blind people’s access to text information on the web. both quantitative and qualitative data were collected to identify the optimal conditions under which the prototype can best serve the blind. significance of study public libraries in the u.s., primar- ily through their main libraries, are providing special services for the visually impaired. currently, the core service is the provision of audio- books. as digital libraries prevail, services for the blind should be online as well, with the the library of congress’s web-braille service as one of the leading examples. however, such services require the use of an expensive braille output device. upon refinement, this prototype would signifi- cantly improve the experience of the visually impaired using online ser- vices. this proto- type can be easily expanded to sup- port graphical dis- plays without any additional devices, making the use of touchable picture books possible for blind users in libraries. prototype development force feedback technology has been used for many years in video games. its use has expanded to other areas such as surgical operations and dan- gerous mechanical processes. this technology was previously applied to gis to solve the problem of ambiguous multicolor displays for multi-variable thematic maps. the same technique was used for this project. the online braille generator translates text on the web into a braille display, letting the user feel the braille dots with a vibrating mouse. the prototype interface was developed using immersion studio (www.immersion.com), javascript, perl/cgi, and active server pages (asp). logitech’s ifeel mouse, inex- pensive at a cost of $ , was used for force feedback output (figure ). the interface has an input text box, which can be filled with any plain text. once it is submitted, the text is instantly translated into braille (fig- ures and ). when the user moves the mouse over each dot on the screen, it vibrates with a given force. while users explore the screen with the vibrating mouse, force feedback dots provide a tactile effect similar to braille displays. in future projects, the manual con- figure . experimental setting information technology and libraries | march information technology and libraries | march version programs will be upgraded to automatic conversion programs with which any texts on the web can be grabbed by their urls and converted into a touchable format for the blind. participants to make this prototype more usable, user studies were conducted in milwaukee, wisconsin, with par- ticipants who are completely blind and read braille. the small sample size—due to the relatively small per- centage of visually impaired peo- ple who are completely blind and can read braille—is comparable to or larger than those found in other research on the blind. the participants came from vari- ous age groups—teens ( ), twenties ( ), thirties ( ), fifties ( ), and sixties ( )—and included females and males. nineteen of the were born blind. participants were recruited at several sites, including the university of wisconsin–milwaukee student accessibility center, public librar- ies with centers for the blind, and nonprofit organizations for the physi- cally impaired. vision teachers in local school districts were also contacted. participants provided valuable i n f o r m a t i o n about the optimal con- ditions for the use of the prototype. this infor- mation will eventually lead to force feedback displays that enable visu- ally impaired people to access the vast amount of information on the web without expensive devices. experimental procedure experiments were conducted in a number of settings, including in the organization’s offices, at the partici- pants’ homes, and at the site of a regional annual meeting for the blind. each ses- sion lasted no more than minutes. participants were asked to try differ- ent interfaces of force feedback braille out- puts with various dot sizes and magnitudes of force. they used a tactile mouse on a note- book computer; after exploring every option, they were asked to select the most comfort- able settings for their sense of touch, including what size the dots should be, how strong the force should be, what kind of force feedback should be used (vibration or friction), and their general opinions of the prototype (see figure ). interviews accompanied the experiments so that both quantita- tive and qualitative data could be col- lected. interviews were transcribed for qualitative data analysis. result even though there were only study participants, a number of issues were clearly identified. it is encouraging to see that all of the participants could identify braille characters using the force feedback mouse with the guidance of the researcher. all the participants agreed that this prototype would be useful with training. the participants pre- ferred the largest dot size ( pixels in diam- figure . touchable braille input screen figure . touchable braille output screen figure . inexpensive force feedback mouse article title | author touchable online braille generator | jeong eter) and the strongest force possible for maximum perception of the force feedback effect. however, the prototype was less attractive to the participants than the currently dominant voice synthesizer software. at least two participants mentioned that their current braille pads fulfill their needs. it seems that they are not motivated to invest their time and effort in a new device. when a potential graphical dis- play application was introduced at the end of a session, the participants became more receptive. at this time there is no practical solution for the visually impaired to feel graphics on computers. experimental devices are available, but they are either quite expensive or still in the research phase. the blind participants also suggested that this graphical proto- type could be used for geometry and geography easily and effectively. discussion blind people’s navigation by mouse because blind people do not use a mouse for computing, using the force feedback mouse itself was a challenge for the study participants. a sighted person uses a mouse with both hand and eye, moving the mouse while watching the mouse cursor on the screen. for the blind it is difficult to identify the mouse’s position. the direction of movement and the dis- tance between two points are difficult to grasp. due to the lack of guidance, the blind encounter difficulties in moving the mouse in a straight line. these issues hinder the effectiveness of force feedback displays for the blind. however, this issue does not only affect the blind. some sighted people, especially older adults, can- not move a mouse easily. one pos- sible solution may be to develop guardrails to help blind people to dif- ferentiate relevant areas of the screen from irrelevant ones. due to their inexperience in using a mouse, the participants held the mouse too firmly to move it or to feel the force feedback. the only participant to use the mouse success- fully was a college student who is music major with years of piano playing experience. this implies that a significant learning session will be required to allow blind people to use the mouse freely. ignorance or suppression of graphical information need even though the participants were more excited about the potential graphical displays, blind people’s graphical information needs are lim- ited. it is possible that their graphi- cal information needs are ignored or suppressed based on their life- time experiences. they tend to resort to braille and, more recently, voice synthesizers instead of graphi- cal displays. this finding suggests the importance of studying the real information needs of the blind or visually impaired rather than the sighted researchers’ expectations of those needs. more research needed with sound because the blind already use sound, particularly voice synthesiz- ers, more sound applications should be researched. for example, audio games have the potential to help blind children learn some skills in the same way that video games teach certain skills to sighted children. audio games also provide a broader research area for future studies. conclusion numerous devices have been devel- oped to improve blind or visually impaired people’s access to informa- tion, including information on the internet. however, such devices are quite expensive or limited in flex- ibility and mainly work in text-only environments. there is no suitable graphic display for the blind, except the laboratory level’s expensive and bulky pin-based external devices. this new prototype uses estab- lished force feedback technology with a minimal cost to existing pcs. it functions for both text and graph- ics. the final products derived from this study can be used for many pur- poses nationally and internationally. information on the web can be deliv- ered to the visually impaired without expensive devices. this touchable braille also lets deaf-blind people, who cannot use screen reader soft- ware, access information on the web, and it can help people learn braille. the application of this force feed- back prototype to image displays has exciting and enormous potential because currently there is no practical, usable method for the blind to access images. for example, blind children are still using handmade -d picture books that are labor-intensive and time-consuming to produce. with this prototype, children’s books can be delivered easily to blind children, who will touch the books’ images via the force feedback mouse. maps of local, state, national, or international interests can be delivered to the blind as well. this prototype will help to add yet another sense—touch—to already blossoming visual and auditory digi- tal libraries. through force feedback technology, new multimodal digi- tal libraries will be accessible to the world. acknowledgement this research was supported by a diversity research grant from the american library association in . information technology and libraries | march information technology and libraries | march references and notes . wooseob jeong and myke gluck, “multimodal geographic information systems: adding haptic and auditory display,” journal of the american society for information science and technology , no. ( ): – . . wooseob jeong, “touchable online braille generator,” in proceedings of the th international acm sigaccess conference on computers and accessibility (new york: acm press, ), – . . jack m. loomis and susan j. leder- man, “tactual perception,” in handbook of perception and human performance, ed. k. r. boff, l. kaufman and j. p. thomas (new york: john wiley & sons, ), vol. , chap. , – . . r. dan jacobson, robert kitchen, and reginald golledge, “multimodal vir- tual reality for presenting geographic information,” in virtual reality in geog- raphy, ed. p. fisher and d. unwin (new york: taylor & francis, ), – . . j. kenneth salisbury and man- dayam a. srinivasan, “phantom-based haptic interaction with virtual objects,” ieee computer graphics and applications , no. ( ): – . . christopher ramstein, “combining haptic and braille technologies: design issues and pilot study,” in proceedings of the nd annual acm conference on assis- tive technologies (new york: acm press, ), – . . a. hardwick, s. furner, and j. rush, “tactile access for blind people to virtual reality on the world wide web,” iee colloquium on developments in tactile displays , no. : / – / . . timothy miller and robert zeleznik, “the design of d haptic wid- gets,” in proceedings of the symposium on interactive d graphics (new york: acm press, ), – . . r. dan jacobson, “geographic visualization with little or no sight: an interactive gis for visually impaired people (paper submitted to aag-gis spe- cialty group student paper competition). . r. brent gillespie and others, “the virtual teacher” in proceedings of asme dynamic systems and control division (new york: asme, ), vol. , – . . noshira a. langrana and others, “human performance using virtual real- ity tumor palpation simulation,” com- puter & graphics , no. ( ): – . . c. colwell and others, “haptic vir- tual reality for blind computer users,” in proceedings of the third annual acm conference on assistive technologies (new york: acm press, ), – . . wooseob jeong, “exploratory user study of haptic and auditory display for multimodal geographic information systems,” in chi’ extended abstracts on human factors in computing systems (new york: acm press, ), – . . ramstein, “combining haptic and braille technologies.” . wai yu and stephen brewster, “multimodal technologies: multimodal virtual reality versus printed medium in visualization for blind people,” in proceedings of the th international acm conference on assistive technologies (new york: acm press, ), – . . d. tzovaras and others, “multi- modal technologies: design and imple- mentation of virtual environments training of the visually impaired,” in proceedings of the th international acm conference on assistive technologies (new york: acm press, ), – . . r. ramloll and others, “construct- ing sonified haptic line graphs for the blind student: first steps,” in proceedings of the th international acm conference on assistive technologies (new york: acm press, ), – . . v. kathlene emery and others, “toward achieving universal usability for older adults through multimodal feedback,” in proceedings of the con- ference on universal usability (new york: acm press, ), – . . julie a. jacko and others, “older adults and visual impairment: what do exposure times and accuracy tell us about performance gains associated with multimodal feedback?” in proceed- ings of the sigchi conference on human factors in computing systems (new york: acm press, ), – . . seongzoo jeong, naoki hashimoto, and sato makoto, “a novel interaction system with force feedback between real and virtual humans,” in proceedings of the acm sigchi international con- ference on advances in computer entertain- ment technology (new york: acm press, ), – . . jeong and gluck, “multimodal geographic information systems”; and wooseob jeong, “multimodal trivariate thematic maps with auditory and hap- tic display” (paper contributed to asist , charlotte, north carolina, october –november , ). information technology and libraries | december editorial: farewell and thank you john webb this issue of information technology and libraries (ital), december , marks the end of my term as editor. it has been an honor and a privilege to serve the lita membership and ital readership for the past three years. it has been one of the highlights of my professional career. editing a quarterly print journal in the field of information technology is an interesting experience. my deadlines for the submission of copy for an issue are approximately three and a half months prior to the beginning of the month in which the issue is pub­ lished; for example, my deadline for the submission of this issue to ala production services was august . therefore, most articles that can appear in an issue were accepted in final form at least five months before they were published. some are older; one was a baby at only four months old. when one considers the rate of change in information technologies today, one understands the need for blogs, wikis, lists, and other forms of profes­ sional discourse in our field. what role does ital play in this rapidly changing environment? for one, unlike these newer forms, it is double­blind refereed. published articles run a peer review gauntlet. this is an important distinction, not least to the many lita members who work for aca­ demic institutions. it may be crass to state it so baldly, but publication in ital can help one earn tenure, an old­fashioned fact of life. it is indexed or abstracted in nineteen published sources, not all of them in english. many of its articles appear in various digital repositories and archives, and these also are harvested or indexed or both. in addition, its articles are cataloged in worldcat local. many of lita’s most prominent members—your distinguished peers—have published articles in ital. the journal also serves as a source for the wider dis­ semination of sponsored research, a requirement of most grants. and you can read it on the bus or at the beach (heaven forbid!), in the brightest sunlight, or with a flashlight under the covers (though there are no reports of this ever having been observed). i am amazed at how quickly these three years have passed, though that may be at least as much a function of my advanced age as of the fun and pleasure i have had as editor. certainly, these past three years have hosted some notable landmarks in our history. lita and ital both celebrated their fortieth anniversaries. sadly, the death of one of lita’s founders and ital’s first editor, frederick g. kilgour, on july , , at age ninety­two, was a landmark in the passing of an era. oclc and rlg’s merger, which fred lived to witness, was a landmark of a different sort—one of maturity, we hope. ital is now an electronic as well as a print journal. this conversion has had some rough passages, but i trust these will have been ironed out by the time you read this. when i became editor, i had a number of goals for the journal, which i stated in my first editorial in march . reading that editorial today, i realize that we successfully accomplished the concrete ones that were most important to me then: increasing the number of articles from library and i­school faculty; increasing the number that result from sponsored research; increasing the number that describe any relevant research or cutting­edge advance­ ments; increasing the number of articles with multiple authors; and finding a model for electronic publication of the journal. the accomplishment of the most abstract and ambitious goal, “to make ital a destination journal of excellence for both readers and authors,” only you, the readers and authors, can judge. i thank mary taylor, lita executive director, and her staff for all of the support they provided to me during my term. i owe a debt that i can never repay to all of the staff of ala production services who worked with me these past three years. their patience with my some­ times bumbling ways was award­winning. thank all of you. the lita presidents and other officers and board members were unfailingly supportive, and i thank you all. in the lita organizational structure, the ital editor and the editorial board report to the lita publications committee, and the editor is a member of that body. i thank all of the chairs and other members of that commit­ tee for their support. once more, and sadly for the last time, i thank all of the members of the ital editorial board who served dur­ ing my term for their service and guidance. they perform more than their share of refereeing, but more importantly, as i have written before, they are the junkyard dogs who have kept me under control and prevented my acting on my worst instincts. i say again, you, the lita member­ ship and ital readership, owe them more than you can ever guess. trust me. to marc truitt, ital managing editor and the incom­ ing ital editor for the – volume years, i must say, “thank you, thank you, thank you!” marc and the ala production services staff were responsible for the form, fit, and finish of the journal issues you received in the mail, held in your hands, and read under the covers. finally, most of all, thank you authors whose articles, communications, and tutorials i have had the privilege to publish, and you whose articles have been accepted and await publication. john webb (jwebb@wsu.edu) is a librarian emeritus, washington state university, and editor of information technology and libraries. editorial: farewell and thank you | john webb not only is this the end of my term as editor, but i also have retired. from now on, my only role in the field of library and information technology will be as a user. those of you have seen the movie the graduate probably remember the early scene when benjamin, the dustin hoffman character, receives the single word of advice regarding his future: “plastics.” (i don’t know if that scene is in the novel from which the movie was adapted.) my single word of advice to those of you too young or too ambitious to retire from our field is: “handhelds.” i am surprised that my treo is more valuable to me now in retirement than it was when i was working. (i’m not surprised that my ipod video is, nor that word thinks that treo and ipod are misspellings.) i just wish that more of the web was as easily accessible on my treo as are google maps and almost all of yahoo!. handhelds. trust me. information technology and libraries | december author id box for column layout column title editor enterprise digital asset management (dam) systems are beginning to be explored in higher education, but little information about their implementation issues is avail- able. this article describes the university of michigan’s investigation of managing and retrieving rich media assets in an enterprise dam system. it includes the background of the pilot project and descriptions of its infrastructure and metadata schema. two case studies are summarized—one in healthcare education, and one in teacher education and research. experiences with five significant issues are summarized: privacy, intellec- tual ownership, digital rights management, uncataloged materials backlog, and user interface and integration with other systems. u niversities are producers and repositories of large amounts of intellectual assets. these assets are of various forms: in addition to text materials, such as journal papers, there are theses, performances from per­ forming arts departments, recordings of native speakers of indigenous languages, or videos demonstrating surgical procedures, to name a few. such multimedia materials have not, in general, been available outside the originat­ ing academic department or unit, let alone systematically cataloged or indexed. valuable assets are “lost” by being locked away in individual drawers or hard disks. managing and retrieving multimedia assets are not problems confined to academia. media companies such as broadcast news agencies and movie studios also have faced this problem, leading to their adoption of digital asset management (dam) systems. in brief, dam systems are not only repositories of digital­rich media content and the associated metadata, but also provide management functionalities similar to database manage­ ment systems, including access control. a dam system can “ingest digital assets, store and index assets for easy searching, retrieve assets for use in many environments, and manage the rights associated with those assets.” in summer , the university of michigan (u­m) tv station, umtv, was searching for a video archive solution. that fall, a u­m team visited cnn and experienced a “eureka!” moment. as james hilton, then­associate provost for academic, information, and instructional technology affairs, later wrote, “building a digital asset management into the infrastructure . . . will be the digital equivalent of bringing indoor plumbing to the campus.” in spring , an enterprise dam system was considered for inclusion in the university infrastruc­ ture. upon completion of a limited proof­of­concept project, a cross­campus team developed the request for proposals (rfp) for the dams living lab, which was issued in july and subsequently awarded to ibm and ancept. in august , hardware and software installation began in the living lab. by , the project changed its name to bluestream to appeal to the grow­ ing mainstream user base. six academic and two support units agreed to partner in the pilot: ■ school of education ■ school of dentistry ■ college of literature, science, and the arts ■ school of nursing ■ school of pharmacy ■ school of social work ■ information technology central services ■ university libraries the academic units were asked to provide typical and unusual digital media assets to be included in the living lab pilot. the pilot focused on rich media, so the preferred types of assets were digital video, images, and other multimedia delivered over the web. the living lab pilot was designed to address four key questions: ■ how to create a robust infrastructure to process, manage, store, and publish digital rich media assets and their associated metadata. ■ how to build an environment where assets are eas­ ily searched, shared, edited, and repurposed in the academic model. ■ how to streamline the workflow required to create new works with digital rich media assets. ■ how to provide a campuswide platform for future application of rights declaration techniques (or other ip tools) to existing assets. this article describes the challenges encountered during the research­and­development phase of the u­m enterprise dam system project known as the living lab. the project has now ended, and the implemented project is known as bluestream. enterprise digital asset management system pilot: lessons learned yong-mi kim, judy ahronheim, kara suzuka, louis e. king, dan bruell, ron miller, and lynn johnson yong-mi kim (kimym@umich.edu) is carat-rackham fellow , school of information; judy ahronheim (jaheim@umich .edu) is metadata specialist, university libraries; kara suzuka (ksuzuka@umich.edu) is assistant research scientist, school of education; louis e. king (leking@umich.edu) is managing producer, digital media commons; dan bruell (danlbee@umich .edu) is director, school of dentistry; ron miller (ronalan@umich .edu) is multimedia services position lead, school of education; and lynn johnson (lynjohns@umich.edu) is associate professor, school of dentistry, university of michigan, ann arbor. article title | author enterprise dam system pilot | kim, ahronheim, suzuka, king, bruell, miller, and johnson ■ background of the living lab: u-m enterprise dam system project an enterprise project such as the living lab at u­m can have significant impact on an institution’s teaching and learning activities by allowing all faculty and students easy yet secure access to media assets across the entire campus. such extensive impact can only be obtained by overcoming numerous and varied obstacles and by docu­ menting actual implementation experiences employed to overcome those challenges. enterprise dam system vendors such as stellent, artesia, and canto list clients from many different industry sectors, including gov­ ernment and education, but provide no detailed case studies on their web sites. information regarding the status of enterprise dam system projects and specific issues that arose during implementation is difficult to find. information publicly available for enterprise dam system projects in higher education is usually in the form of white papers or proposals that do not cover the actual implementations. given the high degree of interest and the number of pilot projects announced in recent years, this shortcoming has prompted the writing of this article, which presents the most important lessons learned dur­ ing the first phase of the living lab pilot project with the hope that these experiences will be valuable to other academic institutions considering similar projects. as part of its core mission, u­m strives to meet the teaching and learning needs of the entire campus. thus, the living lab pilot solicited participation from a diverse cross­section of the university’s departments and units with the goal of evaluating the use of varied teaching and learning assets for the system. from the beginning, it was expected that this system would handle assets in many different forms, such as digital video or digitized images, and also accommodate various organizational schemas and metadata for different collections. this sets the u­m enterprise dam system apart from projects that focus on only one type of collection or define a large monolithic metadata schema for all assets. data were gathered through interviews with asset providers, focus groups with potential users, and a review of the relevant literature. a number of barriers were identified during the pilot’s first phase. while there were some technical barriers, the most signifi­ cant barriers were cultural and organizational ones for which technical solutions were not clear. perhaps the most significant cultural divide was between the culture of academia and the culture of the commercial sector. cultural and organizational assumptions from com­ mercial business practices were embedded in the design of the products initially used in the living lab imple­ mentation. thus, an additional implementation chal­ lenge was determining which issues should be resolved through technical means, and which should be solved by changing the academic culture. this is expected to be an ongoing challenge. ■ architecture (building the infrastructure) an enterprise dam system in an academic community such as u­m needs to support a wide variety of services in order to meet the numerous and varied teaching, research, service, and administrative functions. figure illustrates the services that are provided by an enterprise dam system and concurrently demonstrates its com­ plexity. the left column, process, lists a few of the media processes that various producers will use prepare their media and subsequent ingestion into the enterprise dam system; the middle column, manage, demonstrates the various functions of the enterprise dam system; while the third column, publish, lists a subset of the publishing venues for the media. because an enterprise dam system supports a variety of rich media, a number of software tools and workflows are required. figure illustrates this complexity and describes the architecture and workflow used to add a video segment. the organization of figure parallels that of figure . the left column, process, indicates that flip factory by telestream is used to convert digital video from the original codec to one that can be used for play­ back. in addition, videologger by virage uses media analysis algorithms to extract key frames and time codes created by louis e. king, © regents of the university of michigan figure . component services of the living lab information technology and libraries | december information technology and libraries | december from the video as well as to convert the speech­to­text for easy searching. the middle column, manage, illustrates tools from ibm that help create rich media as well as tools from stellent, such as its ancept media server (ams), that store and index the rich media assets. the third column, publish, illustrates two examples of how these digital video assets could be made available to the end user. one strategy is as a real video stream using real network’s helix server, and the other as a quicktime video stream using ibm’s videocharger. a thorough discussion of all of the software and hardware that make up u­m’s dam system is beyond the scope of this article. however, a list of the software components with links to their associated web sites is provided in figure . from the beginning the living lab pilot aimed for a diverse collection of assets to promote resource discovery and sharing across the university. figure illustrates how the living lab is expected to fit into the varied publishing venues that comprise the campus teaching and learning infrastructure. existing storage and network infrastruc­ tures are used to deliver media assets to various software systems on campus. the living lab is used to streamline the cataloging, searching, and retrieving processes encoun­ tered during academic teaching and research activities. the following example describes how the enterprise dam system fits into the future campus cyberinfrastruc­ ture. a faculty member in the school of music is a jazz composer. one of her compositions is digitally stored in the enterprise dam system along with the associated metadata (cataloging information) that will allow the piece to be found during a search. that single audio file is then found, accessed, and used by five unique publish­ ing venues—the course web site, the university web site, a radio broadcast, the music store, and the library archive. the faculty member uses the piece in her jazz interpreta­ tion course and thus includes a link to the composition on her sakai course web site. when she receives an award, the u­m issues a press release on the u­m web site that includes a link to an audio sample. concurrently, michigan radio uses the enterprise dam system to find the piece for a radio interview with her that includes an audio segment. her performance is published by block m records, u­m’s web­based recording label, and, lastly, the library permanently stores the valuable piece in its institutional archive, deep blue. ■ metadata (managing assets within the academic model) the vision for enterprise dam at u­m is for digital assets to not only be stored in a secure repository, but also be findable, accessible, and usable by the appropriate persons in the university community in their academic endeavors. information about these assets, or metadata, is a crucial component of fulfilling this vision. an important created by louis e. king, © regents of the university of michigan figure . the living lab architecture north american systems ancept media server www.nasi.com/ancept.php ibm content manager www- .ibm.com/software/data/cm/cmgr/mp/ telestream flip factory www.telestream.net/products/flipfactory.htm virage videologger www.virage.com/content/products/index.en.html ibm video charger www- .ibm.com/software/data/videocharger/ real networks helix server www.realnetworks.com/products/media_delivery. html apple quicktime streaming server www.apple.com/quicktime/streamingserver/ handmade software image alchemy www.handmadesw.com/products/image_alchemy. htm figure . software used in the living lab article title | author enterprise dam system pilot | kim, ahronheim, suzuka, king, bruell, miller, and johnson question that arises is, “what kind of metadata should be required for the assets in the living lab?” to help answer this question, potential asset provid­ ers were interviewed regarding their current approach to metadata, such as if they used a particular schema and how well it met their purposes. not surprisingly, asset providers had widely varied metadata implementations. while the assets intended for the living lab pilot all had some metadata, the scope and granularity varied greatly. metadata storage and access methods also varied, ranging from databases implemented using commercial database products and providing web front­ends, to a combination of paper and spreadsheet records that had to be consulted together to locate a particular asset. the assets to be used in the living lab pilot consisted primarily of high­ and low­resolution digital images and digitized video. these interviews also generated a number of requirements for any potential living lab metadata schema. it was deter­ mined that the schema should be able to: ■ describe heterogeneous collections at an appropriate level of granularity and detail, allowing for domain­ specific description needs and vocabularies; ■ allow metadata entry by non­specialists; ■ enable searches across multiple subject areas and col­ lections; ■ provide provenance information for the assets; and ■ provide information on authorized uses of the assets for differing classes of users. an examination of the literature showed a general consensus that no single metadata standard could meet the requirements of heterogeneous collections. projects as diverse as pb core and vius at penn state adopted the approach of drawing from multiple existing metadata standards. their approaches differ in that pb core is a combination of selected metadata elements from a num­ ber of standards plus additional elements unique to pb core, while vius opted for a merged superset of all the elements in the standards selected. in interviews with asset providers (usually faculty), cataloging backlog and the lack of personnel for gen­ erating and entering metadata emerged as consistent problems. there was concern that an overly complex or specialized schema would aggravate the cataloging back­ log by making metadata generation time­consuming and cumbersome. budgetary constraints made hiring pro­ fessional metadata creators prohibitive. another aspect of the personnel problem was that adequate descrip­ tion required subject specialists who were, ideally, the resource authors or creators. but subject specialists, while familiar with the resources and the potential audience for them, may not be knowledgeable of how to produce high­quality metadata, such as controlled vocabularies or consistent naming formats. to address these issues, the more simple and straight­ forward indexing process offered by dublin core (dc) was selected as the starting point for the metadata schema in the living lab. dc was originally developed to sup­ port resource discovery of a digital object, with resource authors as metadata creators. dc is a relatively small standard, but is extensible through the use of qualifiers. it has been adopted as a standard by a number of standards organizations, such as iso and ansi. a body of research exists on its use in digital libraries and its efficacy for author­generated metadata, and there are metadata crosswalks between dc and most other metadata stan­ dards. a number of other subject­specific standards were also examined for more specialized description needs and controlled vocabularies: vra core, ims learning resource meta­data specification, and snodent. in the end, the project leaders elected to adopt a rather novel approach to metadata by not defining one metadata schema for all assets. by taking advantage of the power of multiple approaches (for example, pb core for mix­and­ match, and vius for a merged superset) each collection can have its own schema as long as it contains the ele­ ments of a more general, lowest­common­denominator schema. this overall schema, um_core, was defined based on dc. the elements are prefixed with dc or um to specify the schema origin. um_publisher and um_alternatepublisher identify who should be contacted about problems or ques­ tions regarding that particular asset. um_secondarysubject is a cross­collection subject classification schema devel­ created by louis e. king, © regents of the university of michigan figure . the enterprise dam system as the future campus infra- structure for academic venues information technology and libraries | december information technology and libraries | december oped by the u­m libraries, and helps map the asset into the context of the university. in adopting such an approach to metadata, metadata creation is seen not as a one­shot process, but a collaborative and iterative one. for example, on initial ingestion into the living lab, the only metadata entered for an image may be dc_title, dc_date, and um_publisher. additional meta­ data may be entered as users discover and use the asset, or as input from a subject specialist becomes available. the discussion so far has focused on metadata pro­ duced with human intervention. a number of metadata elements can be obtained from the digital objects through the use of software. in an enterprise dam system, this is referred to as automatically generated metadata and is what can be directly obtained from a computer file such as file name, file size, and file format. this type of metadata is expected to play a larger role as an increasing propor­ tion of assets will be born digital and come accompanied by a rich set of embedded metadata. for example, images or video produced by current digital cameras contain exchangeable image file format (exif) metadata, which include such information as image size, date produced, and camera model used. when available, the living lab presents automatically generated metadata to the user in addition to the elements in um_core. thus, asset metadata in the living lab can be pro­ duced in two ways: automatically generated through a tool such as virage videologger in the case of video, or entered by hand through the current dam system inter­ face. in addition, if metadata already exist in a database format, such as filemaker, this can be imported once the appropriate mappings are defined. videologger, a video analysis tool for digital video files, can extract video key frames, add closed captions, determine whether the audio is speech or music, convert speech to text, and identify (through facial recognition) the speaker(s). these capabilities allow for more sophis­ ticated searching of video assets compared to the cur­ rent capabilities of search engines such as google. some degree of content­based searching can now be done, as opposed to searching that relies on the title and other textual description provided separately from the video itself. for the pilot, particular interest was expressed in the speech recognition capability of videologger. videologger generates a time­coded text of spoken key­ words with to percent accuracy. the result is not nearly accurate enough to generate a transcript, but does indeed provide robust data for searching the content of video. given the diversity of assets in the living lab, it is clear that the university can utilize low­cost keyword analysis to enhance search granularity as well as the more expensive, fully accurate hand­processed transcript. ■ workflow examples two instructional challenges demonstrate how an enter­ prise digital asset management system can provide a solution to instructional dilemmas and how a unique workflow needs to be created for each situation. the chal­ lenges related to each project are described. school of dentistry the educational dilemma the u­m school of dentistry uses standardized patient instructors (spis) to assess students’ abilities to interact with patients. carefully trained actors play carefully scripted patient roles. dental students interview the patients, read their records, and make decisions about the patients’ care, all in a few minutes (see figure ). each session is video recorded. currently, spis grade each student on predeter­ mined criteria, and the video recording is only used if a student contests the spis’ grade. ideally, a dental educator should review each recording and also grade each student. however, the u­m class size of dental students causes a recording­based grading process to be prohibitively expensive in terms of personnel time. in addition, the use of digital videotape makes it difficult for the recorded sessions to be made available to the students. because the tapes are part of the student’s record, they cannot be checked out. if a student wants to review a tape, she or he must make an appointment and review it in a supervised setting. living lab solution the u­m school of dentistry’s living lab pilot attempted simultaneously to improve the spi program and lower the cost of faculty grading spi sessions through three goals: dc_title dc_creator dc_subject um_secondarysubject dc_description dc_publisher dc_contributor dc_date dc_type dc_format dc_identifier dc_source dc_language dc_relation dc_coverage dc_rights um_publisher um_alternatepublisher figure . the u-m enterprise dam system metadata scheme um_core article title | author enterprise dam system pilot | kim, ahronheim, suzuka, king, bruell, miller, and johnson . use speech­to­text analysis to create an easily searched transcript; . streamline the recording process; and . make the videos available online for student review. each of these challenges and the current results are summarized. speech-to-text analysis it was hypothesized that an effective speech­to­text anal­ ysis of the spi session could enable a grader quickly to locate video segments that: ( ) represented student dis­ cussion of specific dental procedures; and ( ) contained student verbalizations of key clinical communication skills. in summer , nine spi sessions were recorded and a comparison between manual transcription and the automated speech­to­text processes was conducted. the transcribed audio track was manually marked up with time­coded reference points and inserted as an annota­ tion track to the video. those same videos also were ana­ lyzed through the video logger speech­to­text service in the living lab, resulting in an automatically generated, time­coded text track. lastly, six keywords were selected that, if spoken by the student, indicated the correct use of either a dental procedure or good communication skills. keyword searches were conducted on both the manual transcription and the speech­to­text analysis. three results were calculated on the key word searches of both versions of all nine recorded sessions. they were: ( ) the number of successful keyword searches; ( ) the number of successful search results that did not actually contain the keywords (false positives); and ( ) the time required to complete the manual transcrip­ tion and text­to­speech analysis of the recordings. the results demonstrated that the speech­to­text analysis matched the manual transcription to percent of the time. also, the speech­to­text process resulted in a false positive less than percent of the time. lastly, the time required to complete the speech­to­text analysis of a session was two minutes, while the average time required to complete a manual transcription of the same session was minutes. while not perfect, the results are encouraging that manually transcribing the audio is no longer necessary. improvements are being made to the clinical environment and microphones so that a higher­quality recording is obtained. it is anticipated that those changes combined with improved software will improve the results of the speech­to­text analysis sufficiently so that automated keyword searches can be conducted for grading purposes. streamlining the recording process scale is a significant challenge to capturing spi inter­ actions in a short amount of time. two to three weeks are required for the entire class of students to complete a series of spi experiences, with as a many as four concur­ rent sessions at any given time. in summer , it was decided to record percent of one class. logistically, one camera operator could staff two stations simultane­ ously. the stations had to be physically close enough for a one­person operation, but not so close that audio from the adjacent session was recorded. the optimal distance was about thirty to thirty­five feet of separa­ tion. staggering the start times of each session allowed the camera operator to make sure each was started with optimal settings. since the results of the speech­to­text analysis were linked to the quality of the equipment used, two prosumer minidv cameras with professional quality microphones and tripods also were purchased. student availability an important strength of living lab is the ability to make the assets both protected and accessible. the current itera­ tion does not have an interface for user­created access con­ trol lists (acl), instead they need to be created by a systems administrator. once a systems administrator has created an acl, academic technology support staff can add or subtract people. to satisfy family educational rights and protection act regulations, a separate acl is needed for each student for the spi project. currently, the possibility of including the spi recordings and their associated transcriptions as ele­ ments of an eportfolio is being explored. in the meantime, students can use url references to include these videos and transcripts in such web­based tools as eportfolios and course management systems. discussion as the challenges of improving speech­to­text analysis, recording workflow, and user­created acls are overcome, the spi program will be able to operate at a new and previ­ ously unimagined level. a more objective keyword grad­ ing process can be instituted. students will be easily able to search through and review their sessions at times and locations that are convenient for them. living lab also will allow students to view their eportfolio of spi interactions and witness how they have improved their communica­ tion skills with patients. for the first time in healthcare education, a clinician’s communication skills, such as bedside or chairside manner, will be able to be taught and assessed using objective methods. school of education the challenge of using records of practice for research and professional education classroom documentation plays a significant role in educational research and in the professional education of teachers at the u­m school of education. collections of information technology and libraries | december information technology and libraries | december videos capturing classroom lessons, small­group work, and interviews with students and teachers—as well as other classroom records, such as images of student work, teacher lesson plans, and assessment documents—are basic to much of the research that takes places in the school of education. however, there also is a large and increasing demand to use these records from real class­ rooms for educational purposes at the u­m and beyond, creating rich media materials for helping preservice and practicing teachers learn to see, understand, and engage in important practices of teaching. this desire to create widely distributed educational materials from classroom documentation raises two important challenges: first, there is the important challenge of protecting the identity of children (and, in some cases, teachers); and second, there is the difficult task of ensuring that the classroom records can be easily accessed by individuals who have permission to view and use the records while being inac­ cessible to those without permission. one research and materials development project at the u­m school of education has been exploring the use of living lab to support the critical work of processing classroom records for use in research and in educational materials, and the distribution and protection of class­ room records as they are integrated into teacher educa­ tion lessons and professional development sessions at the u­m and other sites in the united states. the findings and challenges of these efforts are summarized below. processing classroom records the classroom records used in the pilot were processed in three main ways, producing three different types of products: ■ preservation copies are high­quality formats of the classroom records with minimal loss of digital infor­ mation that can be read by modern computers with standard software. these files are given standardized filenames, cleaned of artifacts and minor irregu­ larities, and de­identified (that is, digitally altered to remove any information that could reveal the identity of the students and, in some cases, of the teachers). ■ working copies are lower­quality versions of the preservation copies that are still sufficient for print­ ing or displaying and viewing. trading some degree of quality for smaller file sizes and thus data rates, the working copies are easier for people to use and share. additionally, these files are further devel­ oped to enhance usability: videos are clipped and composited to feature particular episodes; videos also are subtitled, flagged with chapter markers (or other types of coding), and embedded with links for accessing other relevant information; images of stu­ dent and teacher work are organized into multipage pdfs with bookmarks, links, and other navigational aids; and all files are embedded with metadata for aiding their discovery and revealing information about the files and their contents. ■ distribution copies are typically similar in quality to the working copies but are often integrated into other documents or with other content; they are labeled with copyright information and statements about the limitations of use. they are, in many cases, edited for use on a variety of platforms and copy protected in small ways (for example, word and powerpoint files are converted to pdfs). the living lab was found to support this processing of classroom records in two important ways. first, the system allowed for the setup and use of workflows that enabled undergraduate students hired by the project to upload processed files into the system and walk through a series of quality checks, focused on different aspects of the products. so, for example, when checking the preservation copies, one person was assigned to check the preservation copy against the actual artifact to make sure everything was captured adequately and that the resulting digital file was named properly (“quality check ”). another individual was assigned to make sure the content was cleaned up properly and that no identifying information appeared anywhere (“quality check ”). and finally, a third person checked the file against the meta­ data to make sure that all basic information about the file was correct (“quality check ”). files that passed through all checks were organized into collections accessible to project members and others (“organize”). files that failed along the way were sent back to the beginning of the workflow (the “drawing board”), fixed, and checked again (see figure ). figure . a dental student interviewing an spi. article title | author enterprise dam system pilot | kim, ahronheim, suzuka, king, bruell, miller, and johnson second, living lab allowed asset and collection development to be carried out collaboratively and itera­ tively, enabling different individuals to add value in dif­ ferent ways over time. undergraduate students did much of the initial processing and checking of the assets; skilled staff members converted subtitles into speech metadata housed within living lab; and, eventually, project faculty and graduate students will add other types of analytic codes and content specific metadata to the assets. distribution and protection of classroom records in addition to supporting the production of various types of assets and collections, the living lab supported the distribution and protection of classroom records for use in education settings both at u­m and other institutions. for example, almost fifteen hours of classroom videos from a third­grade mathematics class were made acces­ sible to and were used by instructors and students in the college of education at michigan state university. in a different context, approximately ten minutes of classroom video was made available to instructors in mathematics departments at brigham young university, the university of georgia, and the city college of new york to use in courses for elementary teachers. each asset (and its derivatives) housed within living lab has a url that can be embedded within web pages and online course­management systems, allowing for a great deal of flexibility in how and where the assets are pre­ sented and used. at the same time, each call to the server is checked and, when required, users are prompted to authen­ ticate by logging in before any assets are delivered. this has great potential for easily, seamlessly, and safely integrating living lab assets into a variety of web spaces. although this feature has indeed allowed for a great deal of flexibility, there were and continue to be challenges with creating an integrated and seamless experience for school of education students and their instructors. for example, depending on a variety of factors, such as user operating systems and web browser combinations, users might be prompted for multiple logins. additionally, the login for the living lab server can be quite unforgiving, locking out users who fail to login properly in the first few tries and providing limited communication about what has occurred and what needs to be done to correct the situation. discussion during the living lab pilot a number of workflow chal­ lenges were overcome that now allow numerous and varied types of media related to classroom records to be ingested into living lab, and derivatives created. this demonstrates that living lab is ready for complex media challenges associated with instruction. however, the next challenge of delivering easily and smoothly to others still remains. once authentication and authorization is con­ ducted using single sign­on techniques that allow users to access assets securely from living lab through other systems, assets will be able to be incorporated into web­ based materials and used to enhance the instruction of teachers in ways that have yet to be conceived. ■ privacy, intellectual property, and copyright during the course of the pilot, a number of issues emerged. among these were some of the most critical issues that institutions considering embarking on a similar asset man­ agement system need to address. these issues are: ■ privacy; ■ intellectual ownership and author control of materials; ■ digital rights management and copyright; ■ uncataloged materials backlog; and ■ user interface and integration with other campus systems. up to this point, enterprise dam systems had been developed and used primarily by commercial enterprises— for example, cnn and other broadcasting companies. using a product developed by and for the commercial sec­ tor brought to the fore the cultural differences between the academy and the commercial sector (see figure ). the first three issues in the previous list are related to the differing cultures of commercial enterprise and academia. these issues are addressed below. the fourth and fifth issues are addressed in the section “other important issues.” privacy videos of medical procedures can be of tremendous value to students. in their own words, “watching is different from reading about it in a textbook.” but subjects have the right to retract their consent regarding the use of their images or treatment information for educational purposes. this creates a dilemma: if other assets have been cre­ ated using it, do all of them have to be withdrawn? for drawing board → quality check → quality check → quality check → organize figure . living lab workflow information technology and libraries | december information technology and libraries | december example, if a professor included an image from the univer­ sity’s dam system in a classroom powerpoint or keynote presentation, and subsequently included the presentation in the university’s dam system, what is the status of this file if the patient withdraws consent for use of her or his treatment information? when must the patient’s request be fulfilled? can it be done at the end of the semester, or does it need to be completed immediately? if the request must be fulfilled immediately, the faculty member may not have sufficient time to find a comparable replacement. waiting until the end of the semester helps balance patient privacy with teaching needs. in either case, files must be withdrawn from the enterprise dam system and links to those files removed. consent status and asset relationships must be part of the metadata for an asset to handle such situations. consideration must be given to associating a digital copy of all consent forms with the corresponding asset within an enterprise dam system. intellectual ownership and author control of materials authors’ rights, as recognized by the berne convention for the protection of literary and artistic works, have two components. one, the economic right in the work, is what is usually recognized by copyright law in the united states, being a property right that the author of the work can transfer to others through a contract. the other component—the moral rights of the author—is not explicitly acknowledged by copyright law in the united states and thus may escape consideration regarding ownership and use of intellectual property. moral rights include the right to the integrity of the work, and thus come into play in situations where a work is distorted or misrepresented. unlike economic rights, moral rights cannot be transferred and remain with the author. in a university setting, the university may own the economic right for a researcher’s work, in the form of copyright, but the researcher retains moral rights. the following incident illustrates what can happen when only property rights are taken into account. a digital video segment of a medical procedure was being shown as part of a living lab demo at a university it showcase. because the u­m held the copyright for that particular videotape, no problems were foreseen regarding its usage. a faculty member recognized the video as one she had cre­ ated several years ago and expressed great concern that it had been used for such a purpose without her knowledge or consent. the concern arose from the fact that video showed an outdated procedure. while the faculty member continued to use this video in the classroom, she felt this was different from having it available through the living lab. in the classroom, the faculty member alerted students to the outdated practices during the viewing, and she had full control over who viewed it. the faculty member felt she lost this control and additional clarification when the video became available through living lab. that is, her work was now misrepresented and her moral rights as an author were violated. digital rights management and copyright in the academic world, digital rights management (drm) is becoming a necessary component in disseminating intellectual products of all forms. however, at this time there are few standards and no technical drm solution that works for all media on all platforms. therefore, u­m has elected to use social rather than technical means of managing digital rights. the living lab metadata schema provides an element for rights statements, dc_rights. these metadata, combined with education of the univer­ sity community about copyright, fair use, and the highly granular access control and privileges management of the system, provide the community with the knowledge and tools to use the assets ethically. the university can establish rights declarations to use in the dc_rights field as standards are developed and prec­ edent is established in the courts. these declarations may include copyright licenses developed by the university legal counsel as well as those from the creative commons. current solution—access control lists a clear difference between the cultures of commercial enterprises and academia emerged regarding access to assets, administered through acls. an acl specifies commercial dam system model university dam system model assets held centrally federated ownership of assets access, roles, and privileges managed centrally distributed management of access, privileges and roles metadata frameworks— monolithic federated metadata schema agnostic user interface(s) re: privileges, ownership figure . differences between commerical and university uses of a dam system. article title | author enterprise dam system pilot | kim, ahronheim, suzuka, king, bruell, miller, and johnson who is allowed to access an asset and how they can use it. in commercial settings, access to assets is centrally managed, while in academia, with its complex set of intellectual and copyright issues, it is preferable to have them managed by the asset holders. university users repeatedly asked for the ability to define acls for each asset in the living lab. currently, end users and support staff cannot define acls—only system administrators can create them. the middleware for user­defined acls has been fully developed, and the user interface for user­ defined acls will be made available in the next version. this capability is important in the academic envi­ ronment because the composition of group(s) of people requiring access to a particular asset is fluid and can span many organizational boundaries, both within and outside the university. a research group owning a collection of assets may want to restrict access for various reasons, including requirements set forth by an institutional review board (irb, a university group that oversees research projects involving human subjects), or regulations such as the health insurance portability and accountability act of , which addresses patient health information privacy. the research group will want flexible access control, as research group members may collaborate with others inside and outside the university. the original irb approval may specify that confidentiality of the subjects must be maintained, and collected data, such as video or transcripts, can only be viewed by those directly involved in the research project and cannot be browsed by other researchers not involved in the study or the public at large. in another situation, a collection of art images may only be viewed by current students of the institution, thus requiring a different acl. this situation is still open to interpretation. some say patient consent regarding the use of information for instructional purposes cannot be withdrawn for the use of existing information at the home institution. they can only withdraw it for the use of future assets. others may feel that patients can withdraw permission for the use of their patient assets. other important issues uncataloged materials backlog what emerged from interviews and focus groups with content providers was that while there was no lack of assets they would like to see online, a large proportion of these assets had never been cataloged or even sys­ tematically labeled in some form. this finding may be attributed in part to the pilot focusing on existing assets that have previously not been available for widespread sharing—such as the files stored on faculty hard disks and departmental servers—only known to a favored few. owners or creators of these materials had not consciously thought about sharing these materials or making them available to others. librarians, in contrast, have devel­ oped systems and practices to ensure the findability of materials that enter the library. asset owners were more than willing to have the assets placed online, but did not have the time or resources to provide the appropriate metadata. hiring personnel to create the metadata is problematic, as there is a limit to the metadata that can be entered by non­experts, and experts often are scarce and expensive. for example, for a collection of oral pathology images of microscopic slides, a subject expert must provide the diagnoses, stain, magnification, and other information for each image. without these details, merely putting the slides online is of little value, but these metadata cannot be provided by laypeople. collaborative metadata creation, allowing multiple metadata authors and iterations, may be one solution to this problem. a number of studies indicate that both organiza­ tional support and user­friendly metadata creation tools are necessary for resource authors to create high­ quality metadata. some of the backlog may be resolved through development of tools aimed at resource authors. in addition, increased use of digital file formats with embedded metadata may contribute to reducing future backlog by requiring less human involvement in meta­ data creation. faculty need to be taught that metadata raises the value and utility of assets. as they come to understand the essential role metadata plays, they, too, will invest in its creation. user interface and integration with other systems an enterprise dam system has two basic types of uses: by producers and by users. producers tend to be digital media technologists who create the digital assets and ingest them into the enterprise dam system. the users are the faculty, students, and staff who use these digital assets in their teaching, learning, or research. the research and development version of the enter­ prise dam system, living lab, works well for digital asset producers, but not for the users of these digital assets. ingestion and accessing processes are quite complex and are not currently integrated with other campus systems, such as the online library catalog or the sakai­based, campuswide course management sys­ tem, ctools. digital producers who are comfortable with complex systems are able to ingest and access rich media. however, users have to log onto the enterprise dam system and navigate its complex user interface. the level of complexity of accessing the media can cre­ ate a barrier to adoption and use. if the level of complex­ ity for accessing the assets is too high for users, then the system also is too complex to expect users to contribute to the ingestion of digital assets. information technology and libraries | december information technology and libraries | december in both student and faculty focus groups there was concern about the technical skills needed for faculty use of an enterprise dam system in the classroom. ideally faculty should be able to incorporate assets seamlessly from the enterprise dam system to their classroom mate­ rials, such as powerpoint or keynote presentations. then, the presentations created on their computers should dis­ play without glitches on the classroom system. obviously faculty members cannot be expected to troubleshoot in the classroom when display problems occur. if the enterprise dam system is perceived as difficult to use, or as requiring a lot of troubleshooting by the user, this will discourage adoption by the faculty. this creates additional demands on the enterprise dam system, and potential additional it staffing demands for the academic units wanting to promote enterprise dam system use. when a problem is experienced in the classroom, the departmental it support, not the enterprise dam system support team, will be the first to be called. ideally, an enterprise dam system should be linked to the campus it infrastructure such that users or con­ sumers do not interact with the dam system itself, but rather through existing academic tools, such as the library gateway, course management system, or departmental web sites. having to learn a new system could be a sig­ nificant barrier to use for many potential dam system users in academia. ■ conclusions and lessons learned the vision of a dam system that would allow faculty and students easy yet secure access to myriad rich media assets is extremely appealing to members of the academy. conducting the pilot projects revealed numerous techni­ cal and cultural problems to resolve prior to achieving this vision. the authors anticipate that other institutions will need to address these same issues before undertaking their own enterprise dam system. using commercial software developed in academia during the course of the living lab pilot, the differ­ ences between academia and the commercial sector proved to be a significant issue. assumptions about the organizational culture and work methods are built into systems, often in a tacit manner. in the case of the initial iteration of the living lab, these assumptions were those of the corporate world, the primary clients of the commercial providers as well the environment of the developers. u­m project participants, meanwhile, brought their own expectations based on the reality of their work environment in academia. universities do not have a strict hierarchical structure, with each aca­ demic unit and department having a great degree of local control. academia also has a culture of sharing, where teaching or research products are often shared with no payment involved, other than acknowledgment of the source. thus, there was a process of mutual edu­ cation and negotiation regarding what was and was not acceptable in the enterprise dam system implementa­ tion. this difference of cultures first manifested itself with acls. in the initial implementation, an acl could be defined only by a system administrator. this was a showstopper for the u­m participants, who thought that asset providers themselves would be able to define and modify the acl for any particular asset. a centralized solution with a single owner of the assets (the company), which is acceptable in the corporate environment, is not acceptable in a university environment, where each user is consumer and owner. defining who has access to an asset can be a complex problem in academia, since this access is a moving target subject to both departmental and institutional constraints. libraries and librarians the traditional role of libraries is one of preserving and making accessible the intellectual property of all of humanity. with each new advance in information tech­ nology, such as dam systems, the role of libraries and librarians continues to evolve. this pilot highlighted the role and value of librarians skilled in metadata develop­ ment and assignment. without their expertise and early involvement, there would have been no standard method of indexing assets, thus preventing users from finding useful media. also, the project reinforced two reasons for encouraging asset creators to assign metadata at the asset creation point instead of at the archival point. one, this ensures that metadata are assigned when the content expertise is available. it is very difficult for producers to assign metadata retrospectively, and the indexing information may no longer be available at the point of archive. two, metadata assignment at the point of asset creation helps to ensure consistent metadata assignment that lends itself to automated solutions at the time of archiving. thus, while their role in digital asset man­ agement systems continues to evolve, the authors predict that the librarians’ role will evolve around metadata, and that libraries will start to become the archive for digital materials. it is anticipated that librarians will work with technical experts to develop workflows that include the automated metadata assignment to help faculty routinely add existing and new collections of assets to the system. one example of such a role is deep blue at the university of michigan. deep blue is a digital framework for pre­ serving and finding the best scholarly and artistic work produced at the university. article title | author enterprise dam system pilot | kim, ahronheim, suzuka, king, bruell, miller, and johnson production productivity new technical complexities emerge with each new asset collection added to the u­m system. new workflows as well as richer software features continue to be developed to meet newly identified integration and user interface needs. as the living lab experience advances, techni­ cal barriers are eliminated and new workflows auto­ mated. the authors anticipate that, eventually, automated workflows will allow faculty and staff to routinely use digital assets with a minimum of technical expertise, thus decreasing the personnel costs associated with the use of rich media. for the foreseeable future, however, techni­ cally knowledgeable staff will be required to develop these workflows and even complete a significant amount of the work. academic practice the more delicate and challenging issue is educating fac­ ulty on the value and power of digital assets to improve their research and teaching. dam is a new concept to fac­ ulty, and it will only become useful when integrated into their daily teaching and research. this will happen as fac­ ulty members become more knowledgeable and increase their comfort in the use of digital assets. the dental case study demonstrates that an improved student experience can be provided with such an asset management system, while the education case study demonstrates that a com­ plex set of authentic classroom materials can be orga­ nized and ingested for use by others. these case studies are only two examples of the unanticipated outcomes that result from the use of digital assets in education. the authors predict that as more unanticipated and innova­ tive uses of digital assets are discovered, these new uses will, in turn, lead to increased academic productivity—for example, teaching more without increasing the number of faculty, students teaching each other with rich media, small­group work, and project­based learning. the list of possibilities is endless. as the living lab evolved from a research and development project into the implementation project known as bluestream, it has become an actual classroom resource. this article described myriad issues that were addressed so that other institutions can embark on their own enterprise dam systems fully informed about the road ahead. the remaining technical issues can and will be resolved over time. the greatest challenges that remain are being discovered as faculty and students use bluestream to improve teaching, learning, and research activities. the success of bluestream specifically, and enterprise dam systems in general, will be determined by their successes and failures in meeting the needs of faculty and students. ■ acknowledgements the authors recognize that the living lab pilot program was conducted with the support of others. we thank ruxandra­ana iacob for her administrative contributions to the project. we thank both ruxandra­ana iacob and sharon grayden for their assistance with writing this article. thanks to karen dickinson for her encourage­ ment, optimism, and constant support throughout the project. we thank mark fitzgerald for his vision regard­ ing the potential of the school of dentistry spi project and for conducting the original research. the living lab pilot was conducted with support from the university of michigan office of the provost through the carat partnership program, which pro­ vided funding for the pilot, and the carat­rackham fellowship program, which funded the metadata work. references . a. doyle and l. dawson, “current practices in digital asset management,” internet /cni performance archive & retrieval working group, , http://docs.internet .edu/ doclib/draft­internet ­humanities­digital­asset­management­ practices­ .html (accessed feb. , ). . d. z. spicer, p. b. deblois, and the educause current issues committee. “fifth annual educause survey identifies current it issues.” educause quarterly , no. ( ): – . . humanities advanced technology and information insti­ tute (hatii), university of glasgow, and the national initiative for a networked cultural heritage (ninch), “the ninch guide to good practice in the digital representation and man­ agement of cultural heritage materials,” , www.nyu.edu/ its/humanities/ninchguide (accessed july , ). . a. mccord, “overview of digital asset management sys­ tems,” educause evolving technologies committee, sept. , . . james l. hilton, “digital management systems,” edu- cause review , no. ( ): . . james. hilton, “university of michigan digital asset management system,” . http://sitemaker.umich.edu/ bluestream/files/dams_year _campus.ppt (accessed feb. , ). . the university of michigan, “bluestream,” , http:// sitemaker.umich.edu/bluestream (accessed feb. , ). . oracle corp., “stellent universal content management,” , www.stellent.com/en/index.htm (accessed feb. , ); artesia digital media group, “artesia: the open text digital media group,” , www.artesia.com/ (accessed feb. , ); canto, “canto,” , www.canto.com (accessed feb. , ). . r. d. vernon and o. v. riger, “digital asset management: an introduction to key issues,” www.cit.cornell.edu/oit/arch­ init/digassetmgmt.html (accessed sept. , ); yan han, “digital content management: the search for a content man­ agement system,” library hi tech , no. ( ): – ; stan­ ford university libraries and academic information resources, information technology and libraries | december information technology and libraries | december “media preservation: digital preservation,” , http://library. stanford.edu/depts/pres/mediapres/digital.html (accessed july , ). . telestream, “telestream, inc.,” , www.telestream.net/ products/flipfactory.htm (accessed feb. , ). . autonomy, inc., “virage products overview: virage vid­ eologger,” , www.virage.com/content/products/index. en.html (accessed feb. , ). . international business machines corp., “ancept media server: digital asset management solution,” , www.nasi. com/ancept.php (accessed feb. , ). . realnetworks, inc., “realnetworks media servers,” , www.realnetworks.com/products/media_delivery.html (accessed feb. , ); apple, inc., “quicktime streaming server,” , www.apple.com/quicktime/streamingserver (accessed feb. , ); international business machines corp., “db content manager video charger,” , www­ .ibm. com/software/data/videocharger/ (accessed feb. , ). . sakai, “sakai: collaboration and learning environment for education,” , www.sakaiproject.org (accessed feb. , ). . the university of michigan, “michigan radio,” , www.michiganradio.org (accessed feb. , ). . the university of michigan, “block m records,” , www.blockmrecords.org (accessed feb. , ); the univer­ sity of michigan, “deep blue,” , http://deepblue.lib.umich. edu (accessed feb. , ). . e. duval et al., “metadata principles and practicalities,” d-lib magazine , no ( ); a. m. white et al., “pb core— the public broadcasting metadata initiative: progress report,” dublin core conference sept. –oct. , , seattle; j. attig, a. copeland, and m. pelikan, “context and meaning: the challenges of metadata for a digital image library within the university,“ college & research libraries , no. (may ): – . . white et al., “pb core—the public broadcasting meta­ data initiative”; attig, copeland, and pelikan, “context and meaning.” . dublin core metadata initiative, “dublin core metadata initiative,” , http://dublincore.org (accessed feb. , ). . visual resources association, “vra core categories, version . ,” , www.vraweb.org/vracore .htm (accessed feb. , ); louis j. goldberg, et al., “the significance of snodent,” studies in health technology and informatics (aug. ): – ; http://ontology.buffalo.edu/medo/sno­ dent_ .pdf (accessed feb. , ). . autonomy, “virage products overview.” . filemaker, inc., “filemaker,” , www.filemaker.com/ products (accessed feb. , ). . m. fitzgerald et al., “efficacy of speech­to­text technol­ ogy in managing video recorded interactions,” journal of dental research , special issue a ( ): abstract no. . . u.s. department of education, “family educational rights and privacy act ferpa,” , www.ed.gov/policy/ gen/guid/fpco/ferpa/index.html (accessed feb. , ). . g. lorenzo and j. ittelson, “an overview of e­portfolios,” educause learning initiative, , http://educause.edu/ir/ library/pdf/eli .pdf (accessed feb. , ). . microsoft corp., “microsoft office powerpoint ,” , http://office.microsoft.com/en­us/powerpoint/default. aspx (accessed feb. , ); apple, inc., “keynote,” , www.apple.com/iwork/keynote (accessed feb. , ). . world intellectual property organization, “berne con­ vention for the protection of literary and artistic works,” , www.wipo.int/treaties/en/ip/berne/trtdocs_wo .html (accessed feb. , ). . wikimedia foundation, inc., “digital rights manage­ ment,” , http://en.wikipedia.org/wiki/digital_rights_ management (accessed feb. , ). . creative commons, “creative commons,” , http:// creativecommons.org (accessed feb. , ). . wikimedia foundation, inc., “access control list,” , http://en.wikipedia.org/wiki/access_control_list (accessed feb. , ). . the university of michigan, “um institutional review boards,” , www.irb.research.umich.edu (accessed feb. , ). . health insurance portability and accountability act of (hipaa), “centers for medicare and medicaid ser­ vices,” , www.cms.hhs.gov/hipaageninfo/downloads/ hipaalaw.pdf (accessed feb. , ). . j. greenberg et al., “author­generated dublin core meta­ data for web resources: a baseline study in an organization,” journal of digital information , no. ( ), http://journals.tdl. org/jodi/article/view/jodi­ / (accessed nov. , ); a. crystal and j. greenberg, “usability of a metadata creation application for resource authors,” library & information science research , no. ( ): – . . the university of michigan, “ctools,” , https:// ctools.umich.edu/portal (accessed feb. , ). . m. cox et al., descriptive metadata for television (amster­ dam: focal pr., ); michael a. chopey, “planning and imple­ menting a metadata­driven digital repository,” cataloging & classification quarterly , no. / ( ): – . article title | author content-based information retrieval and digital libraries | wan and liu content-based information retrieval and digital libraries this paper discusses the applications and importance of content-based information retrieval technology in digital libraries. it generalizes the process and ana- lyzes current examples in four areas of the technology. content-based information retrieval has been shown to be an effective way to search for the type of multime- dia documents that are increasingly stored in digital libraries. as a good complement to traditional text- based information retrieval technology, content-based information retrieval will be a significant trend for the development of digital libraries. w ith several decades of their development, digital libraries are no longer a myth. in fact, some gen- eral digital libraries such as the national science digital library (nsdl) and the internet public library are widely known and used. the advance of computer technology makes it possible to include a colossal amount of information in various formats in a digital library. in addition to traditional text-based documents such as books and articles, other types of materials—including images, audio, and video—can also be easily digitized and stored. therefore, how to retrieve and present this multimedia information effectively through the interface of a digital library becomes a significant research topic. currently, there are three methods of retrieving infor- mation in a digital library. the first and the easiest way is free browsing. by this means, a user browses through a collection and looks for desired information. the second method—the most popular technique used today—is text- based retrieval. through this method, textual information (full text of text-based documents and/or metadata of multimedia documents) is indexed so that a user can search the digital library by using keywords or controlled terms. the third method is content-based retrieval, which enables a user to search multimedia information in terms of the actual content of image, audio, or video (marques and furht ). some content features that have been studied so far include color, texture, size, shape, motion, and pitch. while some may argue that text-based retrieval tech- niques are good enough to locate desired multimedia information, as long as it is assigned proper metadata or tags, words are not sufficient to describe what is some- times in a human’s mind. imagine a few examples: a patron comes to a public library with a picture of a rare insect. without expertise in entomology, the librarian won’t know where to start if only a text-based informa- tion retrieval system is available. however, with the help of content-based image retrieval, the librarian can upload the digitized image of the insect to an online digital image library of insects, and the system will retrieve similar images with detailed description of this insect. similarly, a patron has a segment of music audio, about which he or she knows nothing but wants to find out more. by using the content-based audio retrieval system, the patron can get similar audio clips with detailed information from a digital music library, and then listen to them to find an exact match. this procedure will be much easier than doing a search on a text-based music search system. it is definitely helpful if a user can search this non-textual information by styles and features. in addition, the advance of the world wide web brings some new challenges to traditional text-based information retrieval. while today’s web-based digital libraries can be accessed around the world, users with different language and cultural backgrounds may not be able to do effective keyword searches of these librar- ies. content-based information retrieval techniques will increase the accessibility of these digital libraries greatly, and this is probably a major reason it has become a hot research area in the past decade. ideally, a content-based information retrieval system can understand the multi- media data semantically, such as its objects and categories to which it belongs. therefore, a user is able to submit semantic queries and retrieve matched results. however, a great difficulty in the current computer technology is to extract high-level or semantic features of multimedia information. most projects still focus on lower-level fea- tures, such as color, texture, and shape. simply put, a typical content-based information retrieval system works in this way: first, for each mul- timedia file in the database, certain feature information (e.g., color, motion, or pitch) is extracted, indexed, and stored. second, when a user composes a query, the feature information of the query is calculated as vectors. finally, the system compares the similarity between the feature vectors of the query and multimedia data, and retrieves the best matching records. if the user is not satisfied with the retrieved records, he or she can refine the search results by selecting the most relevant ones to the search query, and repeat the search with the new information. this process is illustrated in figure . the following sections will examine some exist- ing content-based information retrieval techniques for most common information formats (image, audio, and video) in digital libraries, as well as their limitations and trends. gary (gang) wan (gwan@tamu.edu) is a science librarian and assistant professor, and zao liu (zliu@tamu.edu) is a distance learning librarian and assistant professor at sterling c. evans library, texas a&m university, college station, texas. gary (gang) wan and zao liu information technology and libraries | march information technology and libraries | march ■ content-based image retrieval there have been a large number of different content- based image retrieval (cbir) systems proposed in the last few years, either building on prior work or exploring novel directions. one similarity among these systems is that most perform feature extraction as the first step in the process, obtaining global image features such as color, shape, and texture (datta et al., ). one of the most well-known cbir systems is query by image content (qbic), which was developed by ibm. it uses several different features, including color, sketches, texture, shape, and example images to retrieve images from image and video databases. since its launch in , the qbic model has been employed for quite a few digital libraries or collections. one recent adopter is the state hermitage museum in russia (www.hermitage. ru), which uses qbic for its web-based digital image col- lection. users can find artwork images by selecting colors from a palette or by sketching shapes on a canvas. the user can also refine existing search results by requesting all artwork images with similar visual attributes. the following screenshots demonstrate how a user can do a content-based image search with qbic technology. in figure . , the user chooses a color from the palette and composes the color schema of artwork he or she is looking for. figure . shows the artwork images that match the query schema. another example of digital libraries or collections that have incorporated cbir technology is the national science foundation’s international digital library project (www.memorynet.org), a project that is composed of several image collections. the information retrieval sys- tem for these collections includes both a traditional text-based search engine and a cbir system called simplicity (semantics-sensitive integrated matching for picture libraries) developed by wang et al. ( ) of pennsylvania state university. from the front page of these image collections, a user can choose to display a random group of images (figure . ). below each image is a “similar” button; clicking this allows the user to view a group of images that contain similar objects to the previously selected one (figure . ). by providing feedback to the search engine this way, the user can find images of desired objects without knowing their names or descriptions. simply put, simplicity segments each image into small regions, extracts several features (such as color, figure . the general process of content-based information retrieval figure . . a user query figure . . the search results for this query article title | author content-based information retrieval and digital libraries | wan and liu location, and shape) from these small regions, and clas- sifies these regions into some semantic categories (such as textured/nontextured and graph/photograph). when computing the similarity between the query image and images in the database, all these features will be consid- ered and integrated, and best matching results will be retrieved (wang et al., ). similar applications of cbir technology in digital libraries include the university of california–berkeley’s digital library project (http://bnhm.berkeley.edu), the national stem digital library (ongoing), and virginia tech’s anthropology digital library, etana (ongoing). while these feature-based approaches have been explored over the years, an emerging new research direction in cbir is automatic concept recognition and annotation. ideally, automatic concept recognition and annotation can discover the concepts that an image con- veys and assign a set of metadata to it, thus allowing image search through the use of text. a trusted automatic concept recognition and annotation system can be a good solution for large data sets. however, the semantic gap between computer processors and human brains remains the major challenge in the development of a robust auto- matic concept recognition and annotation system (datta et al., ). a recent example of efforts in this field is li and wang’s alipr (automatic linguistic indexing of pictures—real time, http://alipr.com) project ( ). through a web interface, users are able to search images in several different ways: they may do text searches and provide feedback to the system to find similar images. users may also upload an image, and the system will per- form concept analysis and generate a set of annotations or tags automatically, as shown in figure . the system then retrieves images from the database that are visually similar to the uploaded image. in the process of auto- matic annotation, if the user doesn’t think the tags given by the system are suitable, he or she can input other tags to describe the image. this is also the “training” process for the alipr system. since cbir is the major research area and has the lon- gest history in content-based information retrieval, there are many models, products, and ongoing projects in addi- tion to the above examples. as image collections become a significant part of digital libraries, more attention has been paid to possibilities of providing content-based image search as a complement to existing metadata search. ■ content-based audio retrieval compared with cbir, content-based audio retrieval (cbar) is relatively new, and fewer research projects on it can be found. in general, existing cbar approaches start from the content analysis of audio clips. an example of this content analysis is extracting basic audio elements, such as duration, pitch, amplitude, brightness, and band- figure . . a group of random images in the collection figure . . cbir results figure . alipr’s automatic annotation feature information technology and libraries | march information technology and libraries | march width (wold et al., ). because of the great difficulties in recognizing audio content, research in this area is less mature than that in content-based image and video retrieval. although no cbar system has been found to be implemented by any digital library so far, quite a few projects provide good prototypes or directions. one good example is zhang and kuo’s ( ) research project on audio classification and retrieval. the proto- type system is composed of three stages: coarse-level audio segmentation, fine-level classification, and audio retrieval. in the first stage, audio signals are semantically segmented and classified into several basic types includ- ing speech, music, song, speech with music background, environment sounds, and silence. some physical audio features—such as the energy function, the fundamental frequency, and the spectral peak tracks—are examined in this stage. in the second stage, further classification is conducted for every basic type. features are extracted from the time-frequency representation of audio signals to reveal subtle differences of timbre and pattern among different classes of sounds. based on these differences, the coarse-level segmentation obtained in stage one can be classified to narrower categories. for example, speech can be differentiated into the voices of men, women, and children. finally, in the information retrieval stage, two approaches—query-by-keyword and query-by-exam- ple—are employed. the query-by-keyword approach is more like the traditional text-based search system. the query-by-example approach is similar to content-based image retrieval systems where an image can be searched by color, texture, and histogram, and audio clips can be retrieved with distinct features, such as timbre, pitch, and rhythm. this way, a user may choose from a given list of features, listen to the retrieved samples, and modify the input feature set to get more desired results. zhang and kuo’s prototype is a very typical and classic cbar system. it is relatively mature and can be used by large digital audio libraries. more recently, li et al. ( ) proposed a new feature extraction method particularly for music genre classifica- tion named daubechies wavelet coefficient histograms (dwchs). dwchs capture the local and global informa- tion of music signals simultaneously by computing their histograms. similar to other cbar strategies, this method divides the process of music genre classification into two steps: feature extraction and multi-class classification. the music signal information representing the music is extracted first, and then an algorithm is used to identify the labels from the representation of the music sounds with respect to their features. since the decomposition of audio signal can produce a set of subband signals at different frequencies cor- responding to different characteristics, li et al. ( ) proposed a new methodology, the dwchs algorithm, for feature extraction. with this algorithm, the decomposi- tion of the music signals is obtained at the beginning, and then a histogram of each subband is constructed. hence, the energy for each subband is computed, and the charac- teristics of the music are represented by these subbands. one finding from this research reveals that this methodol- ogy, along with advanced machine learning techniques, has significantly improved accuracy of music genre clas- sification (li et al. ). therefore, this methodology potentially can be used by those digital music libraries widely developed in past several years. ■ content-based video retrieval content-based video retrieval (cbvr) is a more recent research topic than cbir and cbar, partly because the digitization technology for video appeared later than those for image and audio. as digital video websites such as youtube and google video become more popular, how to retrieve desired video clips effectively is a great con- cern. searching by some features of video, such as motion and texture, can be a good complement to the traditional text-based search method. one of the earliest examples is the videoq system developed by chang et al. ( ) of columbia university (www.ctr.columbia.edu/videoq), which allows a user to search video based on a rich set of visual features and spatio-temporal relationships. the video clips in the data- base are stored as mpeg files. through a web interface, the user can formulate a query scene as a collection of objects with different attributes, including motion, shape, color, and texture. once the user has formulated the query, it is sent to a query server, which contains several databases for different content features. on the query server, the similarities between the features of each object specified in the query and those of the objects in the database are com- puted; a list of video clips is then retrieved based on their similarity values. for each of these video clips, key-frames are dynamically extracted from the video database and returned to browser. the matched objects are highlighted in the returned key-frame. the user can interactively view these matched video clips by simply clicking on the key- frame. meanwhile, the video clip corresponding to that key-frame is extracted from the video database (chang et al. ). figures . – . show an example of a visual search through the videoq system. many other cbvr projects also examine these content features and try to find more efficient ways to retrieve data. a recent example is wang et al.’s ( ) project, vferret, a content-based similarity search tool for continu- ous archived video. the vferret system segments video data into clips and extracts both visual and audio features as metadata. then a user can do a metadata search or article title | author content-based information retrieval and digital libraries | wan and liu content-based search to retrieve desired video clips. in the first stage, a simple segmentation method is used to split the archived digital video into five-minute video clips. the system then extracts twenty image frames evenly from each of these five-minute video clips for visual feature extraction. additionally, the system splits the audio channel of each clip into twenty individual fifteen- second segments for further audio feature extraction. in the second stage, both audio and visual features are extracted. for visual features, the color element is used as the content feature. for audio features, audio fea- tures originally used by ellis and lee ( ) to describe audio segments are computed. for each fifteen-second video segment, the visual feature vector extracted from the sample image and the audio feature vector extracted from the corresponding audio segment are combined into a single feature vector. in the information retrieval stage, the user submits a video clip query at first, then its feature vector is computed and compared with that of video clips in the database, and the most similar clips are retrieved (wang et al. ). similar projects in this area include carnegie mellon university’s informedia digital video library (www. informedia.cs.cmu.edu) and muvis of finland’s tampere university of technology (http://muvis.cs.tut.fi/index. html). content-based information retrieval for other digital formats with the advance of digitization technology, the content and formats of digital libraries are much richer than before. they are not limited to text, image, audio, and video. some new formats of digital content are emerging. digital libraries of -d objects are good examples. since -d models have arbitrary topologies and can- not be easily “parameterized” using a standard template as in the case for -d forms (bustos et al. ), content- based -d model retrieval is a more challenging research topic than other multimedia formats discussed earlier. so far, four types of solutions—primitive-based, statis- tics-based, geometry-based, and view-based—have been found (bimbo and pala ). primitive-based solutions represent -d objects with a basic set of parameterized primitive elements. parameters are used to control the shape of each primitive element and to fit each primitive element with a part of the model. with statistics-based approaches, shape descriptions based on statistical mod- figure . . the user composes a query figure . . search results for the sample query information technology and libraries | march information technology and libraries | march els are created and measured. geometry-based methods, however, use geometric properties of the -d object and their measures as global shape descriptors. for view- based solutions, a set of -d views of the model and descriptors of their content are used to represent the -d object shape (bimbo and pala ). another novel example is moustakas et al.’s ( ) project on -d model search using sketches. in the experimental system, the vector of geometrical descrip- tors for each -d model is calculated during the feature extraction stage. in the retrieval stage, a user can ini- tially use one of the sketching interfaces (such as the virtual reality interface or by using an air mouse) to sketch a -d contour of the desired -d object. the -d shape is recognized by the system, and a sample primi- tive is automatically inserted in the scene. next, the user defines other elements that cannot be described by the -d contour, such as the height of the object, and manip- ulates the -d contour until it reaches its target position. the final query is formed after all the primitives are inserted. finally, the system computes the similarities between the query model and each -d model in the database, and renders the best matching records. an online demonstration can be found for a european project specifically designed for a -d digital museum col- lection, sculpteur (www.sculpteurweb.org). from its web-based search interface, a user can choose to do a meta- data search or content-based search for a -d object. the search strategy here is somewhat similar to that in some cbir systems: the user can upload a -d model in vrml formats, then select a search algorithm (such as similar color, texture, etc.) to perform a search within a digital collection of -d models. as -d computer visualization has been widely used in a variety of areas, there are more research projects focusing on the content-based informa- tion retrieval techniques for this new multimedia format. ■ conclusion there is no doubt that content-based information retrieval technology is an emerging trend for digital library development and will be an important comple- ment to the traditional text-based retrieval technology. the ideal cbir system can semantically understand the information in a digital library, and render users the most desirable data. however, the machine understand- ing of semantic information still remains to be a great difficulty. therefore, most current research projects, including those discussed in this paper, deal with the understanding and retrieval of lower-level features or physical features of multimedia content. certainly, as related disciplines such as computer vision and artificial intelligence keep developing, more researches will be done on higher-level feature-based retrieval. in addition, the growing varieties of multimedia content in digital libraries have also brought many new challenges. for instance, -d models now become impor- tant components of many digital libraries and museums. content-based retrieval technology can be a good direc- tion for this type of content, since the shapes of these -d objects are often found more effectively if the user can compose the query visually. new cbir approaches need to be developed for these novel formats. furthermore, most cbir projects today tend to be web-based. by contrast, many project were based on client applications in the s. these web-based cbir tools will have significant influence on digital libraries or repositories, as most of them are also web-based. particularly in the age of web . , some large digital repositories—such as flickr for images and youtube and google video for video—are changing people’s daily lives. the implementation of cbir will be a great benefit to millions of users. since the nature of cbir is to provide better search aids to end users, it is extremely important to focus on the actual user’s needs and how well the user can use these new search tools. it is surprising to find that little usabil- ity testing has been done for most cbir projects. such testing should be incorporated into future cbir research before it is widely adopted. bibliography bimbo, a. and p. pala. . content-based retrieval of -d mod- els. acm transactions on multimedia computing, communica- tions, and applications , no. : – . bustos, b., et al. . feature-based similarity search in -d object databases. acm computing surveys , no. : – . chang, s., et al. ). videoq: an automated content based video search system using visual cues. in proceedings of the th acm international conference on multimedia, e. p. glinert, et al., eds. new york: acm. datta r., et al. . content-based image retrieval: approaches and trends of the new age. in proceedings of the th interna- tional workshop on multimedia information retrieval, in con- junction with acm international conference on multimedia, h. zhang, , j. smith, and q. tian, eds. new york: acm. ellis, d. and k. lee. minimal-impact audio-based personal archives. in proceedings of the st acm workshop on continuous archival and retrieval of personal experiences carpe, j. gem- mell, et al., eds. new york: acm. li, t., et al. . a comparative study on content-based music genre classification. in proceedings of the th annual interna- tional acm sigir conference on research and development in information retrieval, c. clarke, et al., eds. new york: acm. li, j. and j. wang, j. . real-time computerized annotation of pictures. in proceedings of the th annual acm international article title | author content-based information retrieval and digital libraries | wan and liu conference on multimedia, k. nahrstedt, et al., eds. new york: acm. marques, o. and b. furht. . content-based image and video retrieval. norwell, mass: kluwer. moustakas, k., et al. . master-piece: a multimodal (gesture+speech) interface for d model search and retrieval integrated in a virtual assembly application. proceedings of the enterface: – . wang, j., et al. . simplicity: semantics-sensitive integrated matching for picture libraries. ieee trans. pattern analysis and machine intelligence , no. : – . wang, z., et al. . vferret: content-based similarity search tool for continuous archived video. in proceedings of the rd acm workshop on continuous archival and retrival of personal experiences, k. maze et al., eds. new york: acm. wold, e., et al. . content-based classification, search, and retrieval of audio. ieee multimedia , no. : – . zhang, t. and c. kuo. . content-based audio classification and retrieval for audiovisual data parsing. norwell, mass.: kluwer. lita national forum cover lita guides cover lita workshops cover index to advertisers statement of ownership, management, and circulation information technology and libraries, publication no. - , is published quarterly in march, june, september, and december by the library information and technology association, american library association, e. huron st., chicago, illinois - . editor: john webb, librarian emeritus, washington state university libraries, pullman, wa - . annual subscription price, $ . printed in u.s.a. with periodical-class postage paid at chicago, illinois, and other locations. as a nonprofit organization authorized to mail at special rates (dmm section . only), the purpose, function, and nonprofit status for federal income tax purposes have not changed during the preceding twelve months. extent and nature of circulation (average figures denote the average number of copies printed each issue during the preceding twelve months; actual figures denote actual number of copies of single issue published nearest to filing date: june issue). total number of copies printed: average, , ; actual, , . sales through dealers and carriers, street vendors, and counter sales: average, ; actual . paid or requested mail subscriptions: average, , ; actual, , . free distribution (total): average, ; actual, . total distribution: average, , ; actual, , . office use, leftover, unaccounted, spoiled after printing: average, ; actual, . total: average, , ; actual, , . percentage paid: average, . ; actual, . . s t a t e m e n t o f o w n e r s h i p , m a n a g e m e n t , a n d c i r c u l a t i o n ( p s f o r m , s e p t e m b e r ) f i l e d w i t h t h e u n i t e d s t a t e s p o s t o f f i c e p o s t m a s t e r i n c h i c a g o , o c t o b e r , . participatory networks | lankes, silverstein, and nicholson author id box for column layout column title editor the goal of the technology brief is to familiarize library decision-makers with the opportunities and challenges of participatory networks. in order to accomplish this goal the brief is divided into four sections (excluding an over- view and a detailed statement of goal): ■ a conceptual framework for understanding and eval- uating participatory networks; ■ a discussion of key concepts and technologies in par- ticipatory networks drawn primarily from web . and library . ; ■ a merging of the conceptual framework with the tech- nological discussion to present a roadmap for library systems development; and ■ a set of recommendations to foster greater discussion and action on the topic of participatory networks and, more broadly, participatory librarianship. this summary will highlight the discussions in each of these four topics. for consistency, the section numbers and titles from the full brief are used. k nowledge is created through conversation. libraries are in the knowledge business. therefore, libraries are in the conversation business. some of those conversations span millennia, while others only span a few seconds. some of these conversations happen in real time. in some conversations, there is a broadcast of ideas from one author to multiple audiences. some conversa­ tions are sparked by a book, a video, or a web page. some of these conversations are as trivial as directing someone to the bathroom. other conversations center on the foun­ dations of ourselves and our humanity. it may be odd to start a technology brief with such seemingly abstract comments. yet, without this firm, if theoretical, footing, the advent of web . , social net­ working, library . , and participatory networks seems a clutter of new terminology, tools, and acronyms. in fact, as will be discussed, without this conceptual footing, many library functions can seem disconnected, and the field that serves lawyers, doctors, single mothers, and eight­year olds (among others) fragmented. the scale of this technology brief is limited; it is to present library decision­makers with the opportunities and challenges of participatory networks. it is only a single piece of a much larger puzzle that seeks to pres­ ent a cohesive framework for libraries. this framework not only will fit tools such as blogs and wikis into their offerings (where appropriate), but also will show how a more participatory, conversational approach to libraries in general can help libraries better integrate current and future functions. think of this document as an overview or introduction to participatory librarianship. readers will find plenty of examples and definitions of web . and social networking later in this article. however, to jump right into the technology without a larger frame­ work invites the rightful skepticism of a library organiza­ tion that feels constantly buffeted by new technological advances. in any environment with no larger conceptual founding, to measure the importance of an advance in technology or practice selection of any one technology or practice is nearly arbitrary. without a framework, the field becomes open to the influence of personalities and trendy technology. therefore, it is vital to ground any technological, social, or policy conversation into a larger, rooted concept. as susser said, “to practice without theory is to sail an uncharted sea; theory without practice is not to set sail at all.” for this paper, the chart will be conversation theory. the core of this article is in four sections: ■ a conceptual framework for understanding and eval­ uating participatory networks; ■ a discussion of key concepts and technologies in par­ ticipatory networks drawn primarily from web . and library . ; ■ a merging of the conceptual framework with the technological discussion to present a sort of roadmap for library systems development; and ■ a set of recommendations to foster greater discussion and action on the topic of participatory networks and, more broadly, participatory librarianship. it is recommended that the reader follow this order to get the big picture; however, the second section should be a useful primer on the language and concepts of partici­ patory networks. ■ library as a facilitator of conversation let us return to the concept that knowledge is created through conversation. this notion stretches back to socrates and the socratic method. however, the specific foundation for this statement comes from conversation theory, a means of explaining cognition and how people learn. it is not the purpose of this article to provide a r. david lankes (jdlankes@iis.syr.edu) is director and associate professor, joanne silverstein (jlsilver@iis.syr.edu) is research professor, and scott nicholson (scott@scottnicholson.com) is associate professor at the information institute of syracuse, (n.y.) syracuse university’s school of information studies. participatory networks: the library as conversation r. david lankes, joanne silverstein, and scott nicholson information technology and libraries | december information technology and libraries | december detailed description of conversation theory, a task already admirably accomplished by pask. rather, let us use the theory as a structure upon which to hang an exploration of participatory networking and, more broadly, participa­ tory librarianship. the core of conversation theory is simple: people learn through conversation. different communities have different standards for conversations, from the scientific community’s rigorous formalisms, to the religious com­ munity’s embedded meaning in scripture, to the some­ times impenetrable dialect of teens. the point remains, however, that different actors establish meaning through determining common definitions and building upon shared concepts. the library has been a place where we facilitate con­ versations, though often implicitly. the concept of learn­ ing through conversation is evidenced in libraries in such large initiatives as information literacy and teaching criti­ cal thinking skills (using such meta­cognitive approaches as self­questioning), and in the smaller events of book groups, reference interviews, and speaker series. library activities such as building collections of artifacts (the tan­ gible products of conversation) inform scholars’ research through a formal conversation process where ideas are supported with evidence and methods. similarly, pres­ ervation efforts, perhaps of wax cylinders with spoken word content or of ancient maps that embody an ongo­ ing dialogue about the shape and nature of the physical world, seek to save, or at least document, important conversations. common use of the word “conversation” is com­ pletely in accordance with the use of the term in conver­ sation theory. the term is, however, more specifically defined as an act of communication and agreement between a set of agents. so, a conversation can be between two people, two organizations, two countries, or even within an individual. how can a conversation take place within an individual? educators and school librarians may be familiar with the term “metacogni­ tion,” or the act of reflecting on one’s learning. yet, even the most casual reader will be familiar with the concept of debating oneself (“if i go right, i’ll get there faster, but if i go left i can stop by jim’s . . .”). the point is that a conversation is with at least two agents trying to come to an understanding. also note that those two agents can change over time. so, while socrates and plato are dead, the conversation they started about the nature of knowl­ edge and the world is carried forward by new genera­ tions of thinkers—same conversation, different agents. people converse, organizations converse, states con­ verse, societies converse. the requirements, in the terms of conversation theory, are two cognitive systems seek­ ing agreement. the results of these conversations, what pask would call “cognitive entanglements,” are books, videos, and artifacts that either document, expand, or result from conversations. so, while one cannot con­ verse with a book, that book certainly can be a starting point for many conversations within the reader and within a larger community. if the theory is that conversation creates knowledge, the library community has added a corollary: the best knowl­ edge comes from an optimal information environment, one in which the most diverse and complete information is available to the conversant(s). library ethics show an implicit understanding of this corollary in the advocacy of intellectual freedom and unfettered access. libraries seek to create rich environments for knowledge and have taken the stance that they are not in the job of arbitrating the conversations that occur or the appropriateness of the information used to inform those conversations. as will be discussed later, this belief in openness of conversations will have some far­reaching implications for the library collec­ tion and is an ideal that can never truly be met. for now, the reader may take away that conversation theory is very much in line with current and past library practice, and it also shows a clear trajectory for the future. this viewpoint’s value is not just theoretical; it has real consequences and uses. for example, much of library evaluation has been based on numeric counts of tangible outputs: books circulated, collection size, reference transactions, and so on. yet this quantitative approach has been frustrating to many who feel they are count­ ing outcomes but not getting at true impact of library service. librarians may ask themselves, “which num­ bers are important . . . and why?” if libraries focused on conversations, there might be some clarity and cohesion between statistics and other outcomes. suddenly, the number of reference questions can be linked to items cat­ aloged or to circulation numbers . . . they are all markers of the scope and scale of conversations within the library context. this approach might enable the library com­ munity to better identify important conversations and demonstrate direct contributions to these conversations across functions. for example, a school district identifies early literacy as important. there is a discussion about public policy options, new programs, and school goals to achieve greater literacy in k– . the library should be able to track two streams in this conversation. the first is the one libraries are accustomed to counting; that is, the library’s contribution to k– literacy (participation in book talks, children’s events, circulation of children’s books, reference questions, and so on). but the library also can document and demonstrate how it furthered the conversation about children’s literacy in general. it could show the resources provided to community offi­ cials. it could show the literacy pathfinders that were created. the point of this example is that the library is both participant in the conversation (what we do to pro­ mote early literacy) and facilitator of conversation (what we do to promote public discourse). article title | author participatory networks | lankes, silverstein, and nicholson the theoretical discussion leads us to a discussion about the second topic of this technology brief: pragmatic aspects of the knowledge as conversation approach, or a participatory approach, as it will be called. as new technologies are developed and deployed in the current environment of limited resources, there must be some means of evaluating their utility. a technology’s util­ ity is appropriately measured against a given library’s mission, which is, in turn, developed to respond to the needs of the community that library serves. first, how­ ever, let us identify some of the new technologies and describe them briefly. ■ participatory networking, social networks, and web . let us now move from the theoretical to the opera­ tional. the impetus behind this article is the relatively recent emergence of a new group of internet services and capabilities. suddenly, terms such as wiki, blog, mashup, web . , and biblioblogosphere have become commonplace. as with any new wave of technological creation, these terms can seem ambiguous. they also come wrapped in varying amounts of hype. they may all, however, be grouped under the phenomenon of par­ ticipatory networking. while we now have a conceptual framework to evaluate these technologies that support participatory networking (for example, do they further conversa­ tions), we still need to know the basics of the terminol­ ogy and technologies. this section outlines key concepts in the pragmatics of participatory networking. the section after this one will join the theoretical and operational to outline key chal­ lenges and opportunities for the library world. we begin with web . . web . much of what we call participatory networking, at least the technological foundation of it, stems from developments in web . . as with many buzzwords, the exact definition of web . is not clear. it is more an aggregation of concepts that range from software development (loosely coupled application programming interfaces [apis] and the ease of incorporating features across platforms) to abstrac­ tions (the user is the content). what pervades the web . approach is the notion that internet services are increas­ ingly facilitators of conversations. the following sections describe some of the characteristics of web . . web . characteristic: social networks a core concept of web . is that people are the content of sites; that is, a site is not populated with information for users to consume. instead, services are provided to individual users for them to build networks of friends and other groups (professional, recreational, and so on). the content of a site, then, comprises user­provided infor­ mation that attracts new members of an ever­expanding network. examples include: ■ flickr. flickr (www.flickr.com) provides users with free web space to upload images and create photo albums. users then can share these photos with friends or with the public at large. flickr facilitates the creation of shared photo galleries around themes and places. ■ the cheshire public library. the teen book blog (http://cpltbb.wordpress.com) at the cheshire public library offers book reviews created only by the stu­ dents who use the library. ■ memorial hall library. the memorial hall library in andover, massachusetts, offers podcasts of poetry contests in which the content is created by students (www.mhl.org/teens/audio/index.htm). ■ libraries in myspace. myspace searches show that there are myspace sites for hundreds of individual libraries and scores of library groups. alexandrian public library (apl), for example, has established a site at myspace (www.myspace.com/teensatapl). this practice is growing among public libraries and is an attempt to reach out to users in their preferred online environments. in this venue, the more friends a library’s myspace site has, the more successful it may be considered. as of this writing, apl had sev­ enty­five friends and fifteen comments. the brooklyn college library had , friends and comments. web . characteristic: wisdom of crowds there has been some research into the quality of mass decision­making. that research shows how remarkably accurate groups are in their judgments. web . pools large groups of users to comment on decisions. this aggregation of input is facilitated by the ready availabil­ ity of social networking sites. certainly, this approach of community organization and verification of knowledge also has its detractors. many, for example, question the wisdom seen in some entries of wikipedia. yet, recent articles have compared this mass editing process favor­ ably to traditional sources of information, such as the encyclopedia britannica. examples include: ■ ebay. ebay has perhaps the most studied and copied community policing and reputation systems. all buyers and sellers can be rated. the aggregation of many users’ experiences create a feedback score that is equivalent to a group credibility rating (see figure ). these kinds of group feedback systems can now be seen in most major internet retailers. ■ librarything. librarything.com makes book recom­ information technology and libraries | december information technology and libraries | december mendations based on the collective intelligence of all users of the site. the greater the pool of collective intelligence, the more information available to the user for decision­making. ■ the diary project. the diary project library (www. diaryproject.com) is a non­profit organization that encourages teens to write about their day­to­day experiences growing up. the goal of this site is to encourage communication among teens of all cul­ tures and backgrounds, provide peer­to­peer support, stimulate discussion, and generate feedback that can help ease some of the concerns teens encounter along the way and let them know that they are not alone. to that end, the site comprises thousands of entries in twenty­four categories. because of the great number of entries, most youth can find helpful materials. web . characteristic: loosely coupled apis an api provides a set of instructions (messages) that a programmer can use to communicate between applica­ tions. apis allow programmers to incorporate one piece of software they may not be able to directly manipulate (code) into another. for example, google maps has made a public api that allows web page designers to include satellite images into their web pages with little more than a latitude and longitude. apis vary in their ease of integration. loosely coupled apis allow for very easy integration using high­level scripting languages such as javascript . examples include: ■ google maps. google maps displays street or sat­ ellite maps showing markers on specific locations provided by an external source with simple sets of longitudes and latitudes. it becomes extremely easy to create geographic information systems with little knowledge of gis principles. ■ flickr. flickr provides easy means to integrate hosted images into other web pages or applications (as with a google map that shows images taken at a specific location). ■ youtube. youtube (www.youtube.com) provides users with the capability to upload and comment upon video on the internet. it also allows for easy integration of the videos into other web pages and blogs. with a simple line of html code, anyone can access streaming video for their content. web . characteristic: mashups mashups are combinations of apis and data that result in new information resources and services. this ease of incorporation has led to an assumption of a “right to remix.” in the world of open source software and the creative commons, the right to remix refers to a grow­ ing expectation among internet users that they are not limited by the interfaces and uses presented to them by a single organization. examples include: ■ chicagocrime.org. an often­cited example of a mashup is chicagocrime.org, which uses google maps to plot crime data for the city of chicago. users can now see exactly which street corner had the most murders. figure shows a marker at the location of every homicide in chicago from november , , to august , . ■ book burro. book burro (http://bookburro.org/ about.html) “is a web . extension for firefox and flock. when it senses you are looking at a page that contains a book, it will overlay a small panel which when opened lists prices at online bookstores such as amazon, buy, half (and many more) and whether the book is available at your library.” ■ library lookup. the mit library lookup greasemonkey script for firefox (http://libraries. mit.edu/help/lookup.html) searches mit’s barton catalog from an amazon book screen. web . characteristic: permanent betas the concept of a permanent beta is, in part, a realization that no software is ever truly complete so long as the user community is still commenting upon it. for example, google does not release services from beta until it has achieved a sufficient user base, no matter how fixed the underlying source code is. permanent beta also is a design strategy. large applications are broken into smaller constituent parts that can be manipulated sepa­ rately. this allows large applications to be continually figure . a seller’s profile shows a potential buyer the ebay com- munity’s current estimation of a seller’s credibility. article title | author participatory networks | lankes, silverstein, and nicholson developed by a more diverse and distributed community (as in open source). examples include: ■ google labs. google has a site named “google labs” (http://labs.google.com) that puts out company­ generated tools and services. in fact, part of a google employee’s work time is dedicated to creating the resources and tools through personal projects and exploration. these tools and services remain a part of the “lab” until they are finished and have sufficient user bases. projects (see figure ) range from the simple (google suggest, which provides a dropdown box of possible search queries as you being to type your search terms) to the extensive (google maps, which started as a google lab project). ■ mit libraries. the mit libraries are experimenting with new technologies to help make access to informa­ tion easier. the tools below are offered to the public with an appeal for feedback and additional tools, and the there is a permanent address designed just to collect feedback from the beta­phase tools, which include: ■ the new humanities virtual browsery, which highlights new books and incorporates an rss feed, the ability to comment on books, links to book reviews, availability information, and links to other books by the same author. ■ the libx—mit edition (http://libraries.mit. edu/help/libx.html), which is a firefox toolbar that allows users to search the barton catalog, vera, google scholar, the sfx fulltext finder, and other search tools; it embeds links to mit­ only resources in amazon, barnes & noble, google scholar, and nyt book reviews. ■ the dewey research advisor business and economics q&a (http://libraries.mit.edu/help/ dra.html), which provides starting points for specific research questions in the fields of busi­ ness, management, and economics. web . characteristic: software gets better the more people use it an increasing number of web . sites emphasize social networks, where these services gain value only as they gain users. malcolm gladwell recounts this principle and the work of kevin kelly with an earlier telecommunica­ tions network, the network of fax machines connected to the phone system: the first fax machine ever made . . . cost about $ , at retail. but it was worth nothing because there was no other fax machine for it to communicate with. the second fax machine made the first fax machine more valuable, and the third fax made the first two more valuable, and so on. . . . when you buy a fax machine, then, what you are really buying is access to the entire fax network—which is infinitely more valuable than the machine itself. with social networking sites, and all sites that seek to capitalize on user input (reviews, annotations, profiles, etc.), the true value of each site is defined by the number of people it can bring together. a classic example of this characteristic is amazon. amazon sells books and other merchandise, but, in reality, amazon is very much about the marketing of information. amazon gains tremendous value by allowing its users to review and rate items. the more people use amazon and the more they comment, the more visibility these active users gain and the more credibility markers they take on. web . characteristic: folksonomies a folksonomy is a classification system created in a bottom­up fashion with no central coordination. this differs from the deductive approach of such classifica­ tions systems as the dewey decimal system, where the world of ideas is broken into ten nominal classes. it also differs from other means of developing classifications where some central authority determines if a term should be included. in a folksonomy, the members of a group simply attach terms (or tags) to items (such as photos or blog postings), and the aggregate of these terms is seen as the classification. what emerges is a classification scheme that prioritizes common usage (the most­used tags) over semantic clarity (if most people use “car,” but some use “cars,” they are seen as different terms, and the tag “auto­ mobile” has no real relationship within the aggregate classification). examples include: figure : screenshot of chicagocrime.org information technology and libraries | december information technology and libraries | december ■ penntags. penntags (http://tags.library.upenn.edu/ help) is a social bookmarking tool for locating, orga­ nizing, and sharing one’s favorite online resources. members of the penn community can collect and maintain urls, links to journal articles, and records in franklin, the online catalog, and vcat, the online video catalog. once resources are compiled, users can organize them by assigning tags (free­text key­ words) or by grouping them into projects according to specific preferences. penntags also can be used collaboratively, as it acts as a repository of the varied interests and academic pursuits of the penn com­ munity, and a user can find topics and other users related to his or her own favorite online resources. ■ hillsdale teen library. the hillsdale teen library (www.flickr.com/photos/hillsdalelibraryteens) uses flickr to post pictures of events at the hillsdale teen library (figure ). the resulting tag view is repre­ sented in figure . these tags allow users to easily retrieve the images in which they are interested. there are more characteristics of web . , but these give some overall concepts. core new technologies: ajax and web services as we have just discussed, web . is little more than set of related concepts, albeit with a lot of value being currently attached to these concepts. these concepts are supported by two underlying technologies that have facilitated web . development and brought a substantially new (and improved) user experience to the web. the first is ajax, which allows a more desktop­like experience for users. the second is the advent of web services. these technolo­ gies are not necessary for web . concepts, but they have made web . sites much more compelling. ajax ajax stands for asynchronous javascript and xml. it is a set of existing web technologies brought together. at the most basic, ajax allows a browser (the part the user interacts with) and a server (where the data resides) to send data back and forth without needing to refresh the entire web page being worked on. think about the web sites you work with. you click on a link, the browser freezes and waits for the data, then draws it on the screen. early versions of such sites as mapquest would show a map. if you wanted to zoom into the map, you would press a zoom icon and wait while the new map, and the rest of the web page was redrawn. compare this to google maps, where you click in the middle of a map and drag left or right and the map moves dynamically. we are used to this kind of interaction in desktop applications. click and drag has become second nature on the desktop, and ajax is making it second nature on the web, too. another ajax advantage is that it is open and requires only light programming skills. javascript on the client and almost any server­side scripting language (such as active server pages or php) are easily accessible languages. this fact allows for both fast development and easier integration with existing systems. as an example, it should now be easier to bring more interactive web interfaces to existing online catalogs. web services web services allow for software­to­software interactions on the web. using web protocols and xml, applications exchange queries and information in order to facilitate the larger functioning of a system. one example would be a system that uses an isbn number to query multiple online catalogs and commercial vendors for availability (and price) of a book. this simple process might be part of a much larger library catalog that shows users a book and its availability. the point is, that unlike federated search systems such as z . , web services are small. they also tend to be lightweight (that is, limited in what they do), and are aggregated for greater functionality. this is the technological basis for the loosely coupled apis dis­ cussed previously. library . library . is a somewhat diffuse concept. walt crawford, in his extended essay “library . and ‘library . ,’” found sixty­two different (and often contradictory) views and seven distinct definitions of library . . it is no wonder that people are confused. however, it is natural for emerging ideas and groups to function in an environ­ figure : screenshot of current google lab projects article title | author participatory networks | lankes, silverstein, and nicholson ment of high ambiguity. for use in this technology brief, the authors see library . as an attempt to apply web . concepts (and some longstanding beliefs for greater com­ munity involvement) to the purpose of the library. in the words of ormsby, “the purpose of a library is not to . . . showcase new gadgetry . . . ; rather, it is to make possible that instant of insight when all the facts come together in the shape of new knowledge.” in the case of library . , the new gadgetry discussed in the previous section comprises a group of software applications. how the applications are used will determine whether they support ormsby’s “instant of insight.” many libraries and librarians already are pursuing this goal. some, for instance, are using blogs to reach other librarians, their own users (on their own web sites), and potential users (using myspace and other online communities). they are using wikis to deliver reports, teach information literacy, and serve as repositories. one has developed an api that allows wordpress posts to be directly integrated into a library catalog. clearly, the internet and newer tools that empower users seem to be aligned with the library mission. after all, librarians blogging and allowing the catalog to be mashed up can be seen as an extension of current information services. but this abundance of new applications poses a challenge. given the speed with which new tools are invented, librarians may find it difficult to create strate­ gies that include all the desired services that they make possible. for every new application that becomes avail­ able, library administrators must decide whether it can serve the library, how to use it, and how to find additional resources to manage it (for example, “now we can do this. but why should we?”). this problem stems from focusing excessively on the technology. librarians should instead focus on the phenomena made possible by the technology. most important of these phenomena, the library invites participation. as chad and miller state: library . facilitates and encourages a culture of participation, drawing upon the perspectives and con­ tributions of library staff, technology partners and the wider community. library . means harnessing this type of participation so that libraries can benefit from increasingly rich collaborative cataloguing efforts, such as including contributions from partner libraries as well as adding rich enhancements, such as book jackets or movie files, to records from publishers and others. library . is about encouraging and enabling a library’s community of users to participate, contribut­ ing their own views on resources they have used and new ones to which they might wish access. with library . , a library will continue to develop and deploy the rich descriptive standards of the domain, whilst embracing more participative approaches that encourage interaction with and the formation of com­ munities of interest. the carte blanche statement that users participating in the library is “good,” however, is insufficient. library administers must ask, “what is the ultimate goal?” in summary, current initiatives in the library world to bring the tools of web . to the service of library . are exciting and innovative, and, more to the point, they are supportive of the library’s purpose. they may, however, incur costs, such as monitoring blogs and wikis, and cre­ figure : hillsdale teen library figure : hillsdale teen library flickr site information technology and libraries | december information technology and libraries | december ating content and corresponding with users that stretch already inadequate resources even further. ultimately, the value of library . concepts requires us to answer some important questions: will they be used to further knowledge, or will they simply create more work for librarians? what does the next version of library . look like? is its mission the same, and only the tools dif­ ferent? what makes the library different from myspace— simply a legacy? should we incorporate new services into the current library offerings? how do we, as facilitators of conversations, point the way to the next generation of library? it is hoped that some of the concepts in participa­ tory librarianship may answer these questions and help further the innovations of the library . community. participatory networks the authors use the phrase “participatory networking” to encompass the concept of using web . principles and technologies to implement a conversational model within a community (a library, a peer group, the general public, and so on). why not simply adopt social networking, web . , or library . for that matter? let us examine each term’s limitations: ■ social networking: social network sites such as myspace and facebook have certainly captured public attention. they also have proven very popular. in their short life spans, these sites have garnered an immense audience (myspace has been ranked one of the top destination sites on the web) and drawn much atten­ tion from the press. some of that attention, however, has been very negative. myspace, for example, has been typified as a refuge for pedophiles and online predators. even the television show saturday night live has parodied the site for the ease with which users can create false personas and engage in risky online behaviors. to say you are starting a social networking site in your library may draw either enthusiastic support, vehement opposition (“social networking experiment in my library?!”), or simply confused looks. add to the potential negative con­ notations the ambiguity of the term. is a blog a social networking site? is flickr? to compound this confu­ sion, the academic domain of social network theory predates myspace by about a decade. ■ web . : ambiguity also dogs the web . world. for some, it is technology (blogs, ajax, web ser­ vices, and so on). for others, it is simply a buzzword for the current crop of internet sites that survived the burst of the dot­com bubble. in any case, web . certainly implies more than just the inclusion of users in systems. ■ library . : as stated before, the term library . is a vague term used by some as a goad to the library community. further, this term limits the discussion of user­inclusive web services to the library world. while this brief focuses on the library community, it also sees the library community as a potential leader in a much broader field. so, ultimately, the authors propose “participatory net­ working” as a positive term and concept that libraries can use and promote without the confusion and limitations of previous language. the phrase “participatory network” also has a history of prior use that can be built upon. it represents systems of exchange and integration and has long been used in discussions of policy, art, and government. the phrase also has been used to describe online communities that exchange and integrate information. ■ libraries as participatory conversations so where are we? we started with the abstract statement that knowledge is created through conversation. we then looked at the current landscape of technologies that can facilitate these conversations and showed examples of how libraries, other industries, and individuals are using these technologies. in this section we combine the larger framework with the technologies to see how libraries can incorporate participatory networks to further their knowledge mission. participatory librarianship in action let us look specifically at how participatory networks can be used in the library’s role as facilitator of knowledge through conversation. an obvious example is libraries hosting blogs and wikis for their communities, creat­ ing virtual meeting spaces for individuals and groups. indeed, these are increasingly useful functions for librar­ ies. they meet a perceived need in the community and can generate excitement both within the library and in the community. the idea of creating online sites for individu­ als and organizations makes sense for a library, although it is not without difficulties (see the section on challenges and opportunities). libraries also could use freely avail­ able (and increasingly easy to implement) open source software to create library versions of wikipedia (with or without enhanced editorial processes). another way for libraries to offer these services would be through a cooperative or other third­party vendor. such a service easily can be seen as a knowledge management activity capturing and providing local expertise while linking this expertise to that produced at other libraries. another reason for libraries to engage in participatory networking is that one library can more easily collaborate article title | author participatory networks | lankes, silverstein, and nicholson with other libraries in richer dialogues. we currently have systems that connect our online catalogs and share resources through interlibrary loan. these conduits exist and can be used for the transferal of richer data, as has been proved through collaborative virtual reference sys­ tems. in our current systems, as in traditional library practice, when users are referred to other libraries, they are sent out and not brought back. in a participatory library setting, libraries would facilitate a conversation between the user, the community of the local library, and then through the developed conduits, other libraries and their communities. the end result would be a seamless web of libraries where the user can ignore the intrica­ cies of the library’s organization structure and boundar­ ies, and in which the libraries are using the best local resources to meet local needs. bringing libraries seamlessly together to participate in conversations with a single user has another sig­ nificant advantage: the library would make it easy for users to join the conversation regardless of where they are, through the presentation of a single façade. there is, for example, only one google, one amazon, and one wikipedia. why should users have to search from among thousands of libraries to find the conversations they want? participatory networking will be most effective when libraries work together, when the whole is greater than its parts. we currently see elements of the participatory library in the oclc open worldcat project. for example, users searching google may come across a listing provided by oclc. after selecting the entry for the book, the user can then jump to his or her own local library’s information about the book. users do not have to know which library to visit to find a book near them. extending this concept to conversations, one goal of these participatory networks is to make it easier for the user to enter a conversation with the library without having to work to discover their own specific entry points. however, ensuring this effective seamless access to the library will require more than simply adding ele­ ments of participatory networking around the library’s edges. adding services such as blogs and wikis may be seen merely as adjunct to current library offerings. as with any technological advance, scarce resources must be weighed against a desire to incorporate new services. do we expand the collection, improve the web site, or offer blogs to students? a better approach for making these kinds of decisions is to look at the needs of the community served in context with the commonly accepted, core tasks of a library, and see how they can be recast (and enhanced) as conversational, or participatory, tools. in point of fact, every service, patron, and access point is a starting point for a conversation. let’s start with the catalog. if the catalog is a conversation, it is decidedly formal and, more importantly, one way. think of today’s catalog as the educational equivalent of a college lecture. a for­ mal system is used to serve up a series of presentations on a given topic (selected by the user). the presentations are rigid in their construction (marc, aacr , and so on). they follow an abstract model (relevance scores, some­ times alphabetical listings), and provide minimal oppor­ tunities to the receiver of the information to provide feedback or input. they provide no constructive means for the user to improve or shape the conversation. even recent advances in catalog functions (dynamic, graphical visualizations; faceted searching; simple search boxes’ links to non­collection resources) do little more than make the presentation of information more varied. they are still not truly interactive because they do not allow user participation; they do not allow for conversation. to highlight the one­way nature of the catalog, ask a simple question: what happens when the user doesn’t find something? do we assume that the information is there, but that the user is simply incapable of finding it (in which case the catalog presents search tips, refers the patron to an expert librarian who is capable, or offers more information literacy instruction)? do we assume that the information does not exist (refer the patron to interlibrary loan, pass him or her on to a broader search engine)? do we assume that the catalog itself is limited (refer the user to online databases, or other finding aids)? what if we assume that the catalog is just the current place a user is involving in an ongoing conversation —what would that look like? how can such a traditionally rigid system (in concept, more than in any one feature set) be made more participa­ tory? what if the user, finding no relevant information in the catalog, adds either the information or a place­ holder for someone else to fill in the missing information? possibly the user adds information from his or her exper­ tise. however, assuming that most people go to a catalog because they don’t have the information, perhaps the user instead begins a process for adding the information. the user might ask a question using a virtual reference service; at the end of the transaction, the user then has the option to add the question, along with the answer and associ­ ated materials, to the catalog. or perhaps, the user simply leaves the query in the catalog for other patrons to answer, requesting to be notified when a response is posted. in that case, when a new user does a catalog search and runs across the question, he or she can provide an answer. that answer might be a textual entry (or an image, sound, or video), or simply a new query that directs the original questioner or new patrons to existing information in the catalog (user­created see also entries in the catalog). the catalog also can associate conversations with any data point. for example, a user pulls up the record for a book she or he feels might be relevant to an information need she or he is having. this process starts a conver­ sation between that user and the library, its users, and information technology and libraries | december information technology and libraries | december authors of associated works. the user can see comments and ratings associated with this book from not only users of this library, but users of other libraries. also associated is a list of related works and the full audio of a lecture by the author. the user also might be directed to an in­ person or online book group that is reading that book. the point is that the catalog facilitates a conversation as opposed to simply presenting what it “knows” about a topic and then stepping out of the process. the catalog, then, does not simply present information, but instead helps users construct knowledge by allowing the user to participate in a conversation. there are other means of improving (and linking) systems in a conversational library. take the implicit link between the catalog and circulation. of course, these systems have always been linked in that items found in the catalog can be checked out, and checked out items have their status reflected in the catalog. but this kind of state information is a pretty meager offering. imagine using circulation data to improve the actual functionality of the catalog. take the example of a user who is search­ ing the catalog for fictional books on magic. currently, a relevance score between an item’s metadata and the query is computed and then all the items are ranked in a retrieval set. this relevance score can be computed in many ways, but is usually based on the number of times a keyword appears in the record and the placement of that keyword in the metadata record (giving preference to terms appearing in certain marc fields, such as titles). what is missing is the actual, real­world circulation of an item. wouldn’t it make sense, given such an abstract query, to present the user with harry potter first (but not exclusively)? what if we added circulation data to our relevance rankings: how many times this item has been checked out? it turns out that using a simple statistic is amazingly powerful. it is akin to google’s page rank algorithm that presents sites most linked to higher in the results. also, for those worried that users would be flooded with only popular materials, studies show that while these algorithms do change the very top ranked material, the effect quickly fades so that the user can still easily find other materials. another consideration for adjusting a search is to allow the user to tweak the algorithms used to retrieve works. in the example above, a user could turn off the popularity feature. the user also could toggle switches for currency, authority, and other facets of relevancy rankings. the conversational model requires us to rethink the catalog as a dynamic system with data of varying levels of currency and, frankly, quality, coming into and out of the system. in a conversational catalog, there is no reason that some data can’t exist in the catalog for limited dura­ tions (from years to seconds). records of well­groomed physical collections may be a core and durable collection in the catalog, but that is only one of many types of infor­ mation that could exist in the catalog space. furthermore, even this core data can be annotated and linked to (and from) more transient media. so, the user might see a review from a blog as part of a catalog record on one day, but when she or he pulls the record up again in a few days, that review might be absent, the blog writer hav­ ing withdrawn the comment. this is akin to weeding the collection; however, it would happen in a more dynamic fashion than occurs with the content on library shelves. the conversational model also can be used in other areas of the library. what do we digitize? what do we select? what programs do we offer? what do we pre­ serve? the empowered user can participate in answer­ ing all of these questions but does not replace the expert librarian; rather, the user contributes additional and diverse information and commentary. in fact, the catalog scenario just proposed already assumes that the library catalog does more than store metadata. in order for the scenario to work, the catalog must store questions, answers, video, audio—in essence the catalog must be expanded and integrated with other library systems so that a final participatory library system can present a coherent view of the library to patrons. the next section lays out a sort of roadmap for these enhance­ ments and mergers. framework for integration of participatory librarianship as has been noted, participatory networks and libraries as conversations are not brand new concepts sprung from the head of zeus. instead, they are means to integrate past and current innovations and create a viable plan forward. figure provides a sort of road map of how the library might make the transition from current systems to a truly participatory system. it includes current systems, systems under development (such as federated searching), and new concepts (such as the participatory library). it seeks to capture current momentum and push the field forward to a larger view instead of getting bogged down in the intricacies of any one development activity. along the left side of the graph are current library systems. while the terminology may differ from library to library, nearly every system can be found on today’s library web sites. by showing the systems together, the problems of user confusion and library management burden become obvious. users must often navigate these systems based on their needs, and often with little help. should they search the catalogs first, or the databases? isn’t the catalog really just another database? which data­ base do they choose? in our attempts to serve the users better by creating a rich set of resources and services, we have instead complicated their information­seeking lives. as one librarian puts it, “don’t give me one more system i, or my patrons, have to deal with.” article title | author participatory networks | lankes, silverstein, and nicholson from the array of systems on the left side, we can see that libraries have not been doing themselves any favors either. we are maintaining many systems, therefore mak­ ing the calls for yet more systems not only impractical but unwise. the answer is to integrate systems, combining the best of each while discarding the complexity of the whole. the library world is in the midst of doing just that. this section seeks to highlight promising developments in integrating library systems well beyond the library catalog and to highlight not only an ideal endpoint, but also how this ideal system is truly participatory. merging reference and community involvement the functional area furthest along in the integration of participatory librarianship is reference; as reference is most readily recognizable as a conversation, this comes as no surprise. over the last decade, reference services have gone online and have led to shared reference ser­ vices. more importantly, reference done online creates artifacts of reference conversations: electronic files than can be cleaned of personal information and placed in a knowledge base and used as a resource for other users. a new development in reference is the reference blog, in which multiple librarians and other users can be part of a question­answering community with conversations that can live on beyond a single transaction. another functional area of libraries that is already involved with participatory librarianship is community involvement. for decades, public libraries have supported local community groups through meeting spaces. some libraries now are hosting web spaces for local groups. as libraries incorporate participatory technologies into their offerings, they can create virtual places such as discussion forums, wikis, and blogs for these community groups to use. if there are standards for these discussion areas, then groups from different communities also could easily participate in shared boards; this makes sense for groups such as weight watchers or alcoholics anonymous that have local branches and national involvement. in an academic setting, these groups can be student, faculty, or staff organizations or courses. in addition to reference and hosted community con­ versations, the library has been actively creating digi­ tal collections of materials (either through digitization, leasing service from content providers, or capturing the library’s born digital items). parallel to the digital collec­ tion building of library materials is an active attempt to create institutional repositories of faculty papers, teacher lesson plans, organizational documentation, and the like. these services are participatory systems in which col­ lections come from users’ contributions, and they may evolve into digital repositories that include both user­ and librarian­created artifacts. these different conversations can be archived into a single repository, and, if properly planned, the refer­ ence conversations can live alongside, and eventually be intermingled with, the community conversations, and the digital repository (which, after all, though formal, is a community conversation) into a community repository. community repositories allow librarians to be more eas­ ily involved in the conversations of the community and capture important artifacts of these conversations for later use. merging library metadata into an enhanced catalog participatory librarianship can be supported by another functional area of the library: collections. traditionally, the collection comprises books, magazines, and other information resources paid for by the library. electronic resources, such as databases that are leased instead of purchased, make up a large portion of library expen­ ditures. more recently, web­based resources (external feeds and sites) have been selected and added to the virtual collection. several kinds of finding aids are used to locate these information resources. the catalog and databases both contain descriptions of resources and searching interfaces. in order to improve access, libraries include records for databases within the catalog. conversely, federated search­ ing tools combine the records from different databases and could allow the retrieval of both books and articles by com­ bining records from the traditional catalog and databases into one tool. if community­created resources are part of the catalog, then these resources also would be findable alongside other traditional library resources. the tools for describing information resources also can be participatory. in traditional librarianship, the librarians provide metadata that patrons then use to make selections. figure : road map of how the library might make the transition from current systems to a truly participatory system. information technology and libraries | december information technology and libraries | december by examining this use data, recommender systems can be created to help users locate new materials. in participatory networking, patrons will be encouraged to add comments about items. if standards are used for these comments, then they can be shared among libraries to create larger pools of recommendations. as these comments are analyzed, they can be combined with usage databases to create stronger recommender systems to present patrons with additional choices based upon what is being explored. the end result is an enhanced catalog that allows users and libraries to find information regardless of which sys­ tem the information resides in. however, the enhanced catalog is still just that, a catalog. it contains surrogates of digital information and is managed separately from the artifacts themselves. in the case of physical items, this may be all the library systems can manage, but in the case of digital content, there is one more step that needs to be taken. namely, the artificial barrier between catalog (defined as inventory control system) and content (housed in the community repository) must come down. building the participatory library at this point in the evolution of distributed systems into a truly integrated library system, the participatory library, we have two large collections: one of resources, and one of information about the resources. the first collection of digital content, the community repository, is built by the library and its users collaboratively. the second collection, the enhanced catalog, includes metadata, both formal and user­created (such as ratings, commentary, use data, and the like). both the community repository and the enriched catalog are participatory. yet to realize the dream of a seamless system of functionality (seamless to the user and the library), these two systems must be merged, allow­ ing users to find resources and, much more importantly, conversations. furthermore, the users must be able to add to metadata (such as tags to catalog records) and content (such as articles, postings to a wiki, or personal images). the result may be conceived of as a single integrated infor­ mation resource, which, for the purposes of this conversa­ tion, is called the participatory library. users may access the participatory library directly through the library or as a series of services in google, myspace, or their own home pages. the point is that the access to the library takes place at the point of conversa­ tion, not at the point the user realizes he or she needs information from the library. conversations and preservation the conversation model highlights the need for preserva­ tion. aside from simply providing systems that facilitate conversation, libraries serve as the vital community memory. conversations construct knowledge, but some­ one must remember what has already been said and know how to access that dialog. scientific conversations, for example, are built on previous conversations (theories, studies, methods, results, and hypotheses). capturing conversations and playing them back at the right time is essential. this might mean the preservation of artifacts (maps, transcripts, blueprints, photographs), but also it means the increasingly important tasks of capturing the digital dialogs. this highlights the need for institutional repositories (that will later be integrated seamlessly with other library systems, as previously discussed). specifically, web sites, lectures, courseware, and articles must be kept. further, they must be kept in true conversa­ tional repositories that capture the artifacts (the papers), the methods (data, instruments, policy documents), and the process (meeting notes, conversations, presentations, web sites, electronic discussions). they must be kept in information structures that make them readily available as conversations; in other words, users must be able to search for materials and reconstruct a conversation in its entirety from one fragment. being where the conversation is imagine the conversations that are going on in your local library as you read this. imagine the physicist chatting with the gardener, and the trustee talking with the volunteer who is reading the latest best­seller. what knowledge can be gleaned from these novel interac­ tions? can you measure it? can you enhance it? can you capture it? can you recall it when it would be precisely what a user needs? note also that these conversations do not belong solely to the library. the library is only part of the con­ versation. faced with the daunting variety of resources available on the web, many organizations try to become the single point of entry into it. remember that conversa­ tions are varied in their mode, places, and players, and, more importantly, that they are intensely personal. this means that participants need to have ownership in them, and often in their locations as well. this also means that the library, as facilitator, needs to be varied in its modes and access points. in many cases, it is better to either create a personal space in which users may converse, or, increasingly, to be part of someone else’s space. what we can learn from web . ’s mashups is that smaller sets of limited (but easy to access) functionalities lead to greater incorporation of tools into people’s lives. in the chicagocrime–google maps mashup, combining maps from google and chicago crime statistics, it was important for the host of the site to brand the space and shape the interface for his conversation on crime. can your library functions be as easily incorporated into these types of conversations? can a user search your catalog and present the results on his or her web site? the point is that libraries need to be proactive in a new way. instead of article title | author participatory networks | lankes, silverstein, and nicholson the mantra, “be where the user is,” we need to, “be where the conversation is.” it is not enough to be at the users’ desktops; you need to be in their e­mail program, in their myspace pages, in their instant messaging lists, and in their rss feed readers. all of these examples point to a significant mental shift that librarians will need to make in moving from delivering information from a centralized location to delivering information in a decentralized manner where the conversations of users are taking place. the catalog example presented earlier is an example of a centralized place for conversations. what if, instead of only being in a catalog, the same data were split into smaller components and embedded in the user’s browser and e­mail pro­ grams? just as google’s mail system embeds advertising based upon the content of a message, the library could provide links to its resources based upon what a user is working on. by disaggregating the information within its system, the library can deliver just what is needed to a user, provide connections into mashups, and live in the space of the user instead of forcing the user to come to the space of the library. challenges and opportunities there is clearly a host of challenges in incorporating par­ ticipatory networks and a participatory model into the library. this is to be expected when we are dealing with something as fundamental as knowledge and as personal as conversations. we consider four major challenges that must be met by libraries before they can truly get into the business of participatory librarianship. technical there is a rich suite of participatory networking software that libraries can incorporate into their daily operations. implementing a blog, a wiki, or rss feeds these days is not a hard task, and they can easily be used to deliver information about library services and conversations to the user’s space. furthermore, these systems are often tested in very large­scale environments and are, in some cases, the same tools used in large participatory network­ ing sites such as wikipedia and blogger. some of these packages are commercial, but others are open source software. open source software is cheaper, easier to adapt, and, in some cases, more advanced. the downside to open source is that it requires a considerable amount of technical knowledge by the library (but not as much as one might think) and does not come with a technical support hotline. the largest technological impediment, however, may be the currently installed base of software within librar­ ies. integrated library systems have a long history and include a broad range of library functions. legacy code and near monolithic systems have restricted the easy exchange of a diverse set of information. were these sys­ tems written today, they would use modular code and loosely coupled apis and allow customers much more interface customizability. these changes may come to integrated library systems (as customers are demanding it), but it may take years. several libraries are currently attempting to pick apart these integrated systems themselves. often, libraries go to the underlying databases that hold the library metadata or create their own data structures, such as the university of pennsylvania data farm project. once components of this system are exposed, the catalog simply becomes another database that can be federated into new and uni­ fied interfaces. however, such integration requires a great deal of technological expertise. there is an opportunity for integrated library system vendors or large consortial groups such as oclc to move quickly into this space. in the meantime, there is an opportunity for the larger library community. this technology brief was created in response to a perceived need. whether evi­ denced in the library . community or in conversations at lita, libraries are now interested in incorporating new web technologies into their offerings and opera­ tions. the technologies under consideration here pres­ ent platforms for experimentation. rather than setting up thousands of separated experiments, however, the library community should create a participatory net­ work of its own. the technology certainly exists to create a test bed for libraries to set up various combinations of communication technologies (blogs, tagging, wikis), to test new web services against pooled data (catalog data, metadata repositories, and large scale data sets), and even to incorporate new services into the current library offerings (rss feeds, for example). by combining resources (money, time, expertise) in a single, large­scale test bed, libraries not only can get greater impact for the their investments, but can directly experience life as a connected conversation. these connections, if built at the ground level, will then make it easier for the library to come into existence. terminology can be clarified, claims tested, and best practices collaboratively developed, greatly accelerating innovation and dissemination. operational in addition to being in the conversation business, librar­ ies are in the infrastructure business. one of the most powerful aspects of a library is its ability not only to develop a collection of some type of information, but to maintain it over time. sometimes infrastructure can be problematic (as in the case of legacy systems), but more often than not it provides a stable foundation from which to operate. there are many conversations going on that need infrastructure but have none (or little). think of the opportunities in your community for using the web to information technology and libraries | december information technology and libraries | december facilitate a conversation. it might be a researcher want­ ing to disseminate the results of his or her latest study. it might be a community organization seeking funding. it might be a business trying to manage its basic opera­ tional knowledge. the point is that such individuals and community organizations are not in the infrastructure business and could use a partner who is. imagine a local organization coming to the library and, within a few min­ utes, setting up a web site with an rss feed, a blog, and bulletin boards. the library facilitates, but does not own, that individual’s or organization’s conversation. it does form a strong partnership, however, that can be leveraged into resources and support. the true power of participa­ tory networking in libraries is not to give every librarian a blog; it is in giving every community member a blog (and making the librarian a part of the community). in addition, the library can play the role of connecting these conversations to other users when appropriate. participatory libraries allow the concept of com­ munity center (intellectual center, service center, media center, information center, meeting center) to be extended to the web. many public libraries have no problem providing meeting space to local non­profits. why not provide web meeting space in the form of a web site or web conferencing? many academic libraries attempt to capture the scholarly output of their faculties, why not help generate the output with research data stores? the answers to these questions inevitably come back to time and money. however, there is nothing in this brief that says such services have to be free. in fact, the best part­ nerships are formed when all partners are invested in the process. the true problem is that libraries have no idea of how to charge for such services. faculty would be glad to write library support into grants (in the form of web site creation and hosting), but need a dollar figure to include and how long each task will take. many libraries aren’t used to positioning their services on a per item basis, and this makes it difficult to build partnerships. sometimes it is not a lack of money, but a lack of structure to take in money that is the problem. policy as always, it is policy that presents the greatest challenges. the idea of opening the library functions to a greater set of inputs is rife with potential pitfalls. how can libraries use the technologies and concepts of facebook and myspace without being plagued by their problems? how can users truly be made part of the collection without the library being liable for all of their actions? the answers may lie in a seemingly obscure concept: identity management. conversations can range in their mode, topic, and duration. they also can vary in the conversants. the library needs to know a conversant’s status to determine policy (for example, we can only disclose this information to this person), and requires a unique identifier, such as a library card, to uphold it. in traditional libraries, that is the extent of identity management. in a participatory model, distinctions among identi­ ties become complex and graduated, and require us to consider a new approach. this new model, of patrons adding information directly to library systems, is not as radical as it may first appear. we have become very used to the idea of roles and tiered levels of authority in many other settings. most modern computer systems allow for some gradation in user abilities (and responsibilities). online communities have even introduced merit systems, by which continual high­quality contributions to a site equals greater power in the site. think about amazon, wikipedia, even ebay; as users contribute more to the community, they gain status and recognition. from par­ ticipants to editors, from readers to writers, these organi­ zations have seen membership as a sliding scale of trust, and libraries need to adopt this approach in all of their basic systems. we currently do, to a degree, in the form of librarians, paraprofessionals, and other staff. yet even these distinctions tend to be rigid and often class­based, with high walls (such as a master’s degree) between the strata. some of this is imposed by outside organizations (civil service requirements, tenure track, and so on), but a great deal is there by inertia of the field. skillful use of identity management will help librar­ ies avoid the baggage of myspace and facebook. as users grain greater access, greater responsibility, and greater autonomy, libraries need to be more certain of their identities. that is, for a user to do more requires the library to know more. knowing about a user may involve traditional identity verification or tracking an activity trail, whereby intentions can be judged in rela­ tion to actions. these concepts may be expressed as, “the more we know you, the more control you can have in valuable services such as blogging, or the catalog.” the concepts are illustrated in blogger and livejournal, both of which require some level of identity information. in another example, to join livejournal you must be invited, thus the community confers identity. the common theme is that verifying (and building) identity is community­ based. the difference between the library and myspace is that the library works in an established community with traditional norms of identity, whereas myspace is seeking to create a community (where identity is more defined by social connections than actions). both the library and the services mentioned above, however, base their functions and services on identity. ethical as knowledge is developed through conversation, and libraries facilitate this process, libraries have a powerful impact on the knowledge generated. can librarians inter­ fere with and shape conversations? absolutely. should we? we can’t help it. our collections, our reference work, article title | author participatory networks | lankes, silverstein, and nicholson our mere presence will influence conversations. the ques­ tion is, in what ways? by dedicating a library mission to directly align with the needs of a finite community, we are accepting the biases, norms, and priorities of the com­ munity. while a library may seek to expand or change the community, it does so from within. when internet filtering became a requirement for fed­ eral internet funding, public and school libraries could not simply quit, or ignore the fact, because they are agents of their communities. school libraries had to accept filtering with federal funding because their parent organizations, the schools, accepted filtering. we see, from this example, that libraries may shift from facilitating conversations to becoming active conversants, but they are always doing both. thus, the question is not whether the library shapes conversations, but which ones, and how actively? these questions are hardly new to the underlying principles of librarianship. and nothing in the participa­ tory model seeks to change those underlying principles. the participatory model does, however, highlight the fact that those principles shape conversations and have an impact on the community. ■ recommendations the overall recommendation of this article is that librar­ ies must be active participants in the ongoing conversa­ tions about participatory networking. they must do so through action, by modeling appropriate and innovative use of technologies. this must be done at the core of the library, not on the periphery. rather than just adding blogs and photosharing, libraries should adopt the princi­ ples of participation in existing core library technologies, such as the catalog. anything less simply adds stress and stretches scarce resources even further. to complement this broad recommendation, the authors make two specific proposals: expand and deepen the discussion and understanding of participatory net­ works and participatory librarianship, and create a par­ ticipatory library test bed to give librarians needed participatory skills and sustain a standing research agenda in participatory librarianship. as stated in the outset of this document, what you are reading is limited. while it certainly contains the kernel and essence of participatory networks (systems to allow users to be truly part of services) and participatory librar­ ianship (the role of librarianship as facilitators and actors in conversations in general), the focus was on technology and technology changes. already, the ideas contained in this document have been part of an active conversation. the first draft of this document was made available for public comment via a wiki, e­mail, and bulletin boards, and concepts herein presented at conferences and lec­ tures. however, there is now a need to broaden the scope and scale of the conversation. the theoretical founda­ tions of participatory librarianship need to be rigorously presented. the nontechnical components of the ideas (and the marriage of nontechnical to technical) need to be explored. there are curricular implications: how do we prepare participatory librarians? the nature and form of the library and participatory systems need to be dis­ cussed and examined in theoretical, experimental, and operational contexts. in order to do this, the authors propose a series of con­ versations to engage the ideas. these conversations, both in person and virtual, need to be within the profession and across disciplines and industries. the deeper conversa­ tions need to be documented in a series of publications that expand this document for academics and practitioners. the authors feel, however, that the first proposal must be grounded in action. to complement the more abstract exploration of participatory networks and participatory librarianship, there must be an active playground where conversants can experience firsthand the technologies discussed, and then actively shape the tools of partici­ pation. this is the test bed. this test bed would imple­ ment a participatory network of libraries, and provide a common technology platform to host blogs, wikis, discussion boards, rss aggregators, and the like. these shared technologies would be used to experiment with new technologies and to provide real services to librar­ ies. thus, libraries could not only read about blogging applications, they could try them and even roll them out to their community members. as libraries start new com­ munity initiatives, they could rapidly add wikis and rss feeds hosted at the shared test bed. the test bed would also make all software available to the libraries so they could locally implement technologies that have proven themselves. the test bed would provide the open source software and consulting support to implement features locally. the test bed also would develop new metrics and means of evaluating participatory library services for the use of planners and policy makers. a major deliverable of the test bed, however, would be to model innovations in integrated library systems (ils). the test bed would work with libraries and ils vendors to pilot new technologies and specify new stan­ dards to accelerate ils modernization. the point of the test bed is not to create new ilss, but to make it easy to incorporate innovative technologies into vendor and open source ilss. the location and support model of the test bed are open for the library community to determine. certainly, it could be placed in existing library associations or orga­ nizations. however, it would require the host to be seen as neutral in ils issues, and to be capable of supporting a diverse infrastructure over time. the host organiza­ tion also would need to be a nimble organization, able information technology and libraries | december information technology and libraries | december to identify new technical opportunities and implement them quickly. one model that might work is establishing a pooled fund from interested libraries. this pooled fund would support an open source technology infrastructure and a small team of researchers and developers. the team’s activities would be overseen by an advisory panel drawn from contributing members. such a model spreads this investment out into experimentation across a broad col­ laboration and should, ultimately, save libraries time and money. as a result, the time and money that indi­ vidual libraries might spend on isolated or disconnected experiments can be invested in a common effort with greater return. libraries have a chance not only to improve service to their local communities, but to advance the field of par­ ticipatory networks. with their principles, dedication to service, and unique knowledge of infrastructure, libraries are poised not simply to respond to new technologies, but to drive them. by tying technological implementa­ tion, development, and improvement to the mission of facilitating conversations across fields, libraries can gain invaluable visibility and resources. impact and leadership, however, come from a firm and conceptual understanding of libraries’ roles in their communities. the assertion that libraries are an indis­ pensable part of knowledge generation in all sectors pro­ vides a powerful argument to an expanded function of libraries. eventually, blogs, wikis, rss, and ajax all will fade in the continuously dynamic internet environment. however, the concept of participatory networks and con­ versations is durable. ■ acknowledgements the authors would like to thank the following people and groups: ken lavender, for his editing prowess. the doctoral students of ist for providing input on conversation theory: johanna birkland, john d’ignazio, keisuke inoue, jonathan jackson, todd marshall, jeffrey owens, katie parker, david pimentel, michael scialdone, jaime snyder, sarah webb. the students of ist for their tremendous input and for their exploration of the related concept of massive scale librarianship: marcia alden, charles bush, janet chemotti, janet feathers, gabrielle gosselin, ana guimaraes, colleen halpin, katie hayduke, agnes imecs, jennifer kilbury, min­chun ku, todd mccall, virginia payne, joseph ryan, jean van doren, susan yoo. those who commented on the draft, including karen scheider, walt crawford and john buschman, and kathleen de la peña mccook. lita for giving us a forum for feedback. carrie lowe, rick weingarten, and mark bard of ala’s oitp for their feedback and support. the institute staff, including lisa pawlewicz, joan laskowski, and christian o’brien, for logistical support. references and notes . cited in p. hardiker and m. baker, “towards social theory for social work,” handbook of theory for practice teachers in social work, j. lishman, ed. (london: jessica kingsley, ). . g. pask, conversation theory: applications in education and epistemology (new york: elsevier, ). . linda h. bertland, “an overview of research in metacog­ nition: implications for information skills instruction,” school library media quarterly (winter ): – . . pask, conversation theory, . . tim o’reilly, “what is web . : design patterns and business models for the next generation of software,” o’reilly, www.oreillynet.com/pub/a/oreilly/tim/news/ / / / what­is­web­ .html (accessed feb. , ). . j. suroweicki, the wisdom of crowds (new york: double­ day, ). . “wiki’s wild world: researchers should read wikipedia cautiously and amend it enthusiastically,” nature , no. (dec. ): ; www.nature.com/nature/journal/v / n /full/ a.html (accessed feb , ). . google, “google maps api,” www.google.com/apis/ maps (accessed feb. , ). . “java script tutorial,” w schools, www.w schools.com/ js/default.asp (accessed feb. , ). . while the terms in web . are a bit ambiguous, many people confuse the term “mashup” with “remixes.” mashups are combining data and functions (such as mapping), whereas remixes are reusing and combining content only. so combining a song with a piece of video to create a “new” music video would be a remix. mapping all of your videos on a map using youtube to store the videos and google maps to plot them geographically would be a mashup. . for example gmail, a very widely used, web­based email service, but is still considered “beta” by google. . malcolm gladwell, the tipping point: how little things can make a big difference (boston: back bay books, ), . . oclc, “introduction to dewey decimal classification,” www.oclc.org/dewey/versions/ddc print/intro.pdf (accessed feb. , ). . “ajax (programming),” wikipedia, http://en.wikipedia .org/wiki/ajax_(programming) (accessed feb. , ). . “web services activity,” w c, www.w .org/ /ws (accessed feb. , ). . walt crawford, “library . and ‘library . .’ ” cites & insights , no. ( ), http://citesandinsights.info/civ i .pdf (accessed dec. , ). . eric ormsby, “the battle of the book: the research library today,” the new criterion (oct. ): . . ken chad and paul miller, “do libraries matter? the rise of library . : a white paper,” version . , , www.talis .com/downloads/white_papers/dolibrariesmatter.pdf (accessed feb. , ). . slashdot, “myspace # us destination last week,” h t t p : / / s l a s h d o t . o rg / a r t i c l e s / / / / . s h t m l article title | author participatory networks | lankes, silverstein, and nicholson (accessed feb. , ); pete williams, “myspace, facebook attract online predators,” msnbc, www.msnbc.msn.com/ id/ (accessed feb. , ); “the myspace gener­ ation,” businessweek, dec. , , www.businessweek .com/magazine/content/ _ /b .htm (accessed feb. , ). . saturday night live, “sketch: myspace seminar,” nbc, www.nbc.com/saturday_night_live/segments/ .shtml (accessed feb. , ). . c. stohl and g. cheney, “participatory processes/para­ doxical practices,” management communication quarterly , no. ( ): – . . j. zucca, “traces in the clickstream: early work on a management information repository at the university of penn­ sylvania,” information technology and libraries , no. ( ): – . . to be more precise, public and school libraries that accept e­rate funding. information technology and libraries | december author id box for column layout column title editor as public libraries are becoming e-government access points relied on by both patrons and government agen- cies, it is important for libraries to consider the impli- cations of these roles. while providing e-government access serves to reinforce the tremendously important role of public libraries in the united states social infra- structure, it also creates new demands on libraries and opens up significant new opportunities. drawing upon several different strands of research, this paper exam- ines the nexus of public libraries, values, trust, and e-government, focusing on the ways in which the values of librarianship and the trust that communities place in their public libraries reinforce the role of public librar- ies in the provision of e-government. the unique values embraced by public libraries have not only shaped the missions of libraries, they have influenced popular opinion surrounding public libraries and fostered the confidence that communities place in them as a source of trusted information and assistance in finding infor- mation. as public libraries have embraced the provision of internet access, these values and trust have become intertwined with their new social role as a public access point for e-government both in normal information activities and in the most extreme circumstances. this paper explores the intersections of these issues and the relation of the vital e-government role of public libraries to library funding, public policy, library and informa- tion science education, and research initiatives. p ublic libraries have always been valued and trusted institutions within society. due to recent advances in technology and changes in united states society, public libraries now also play a unique and critical role by offering free public internet access. with the increas­ ing reliance on the internet as a key source of news, social capital, and access to government services and information, the free access provided by public librar­ ies is an invaluable resource. as a result, a significant proportion of the u.s. population, including people who have no other means of access, people who need help using computers and the internet, and people who have lower quality access, rely on the internet access and computer help available in public libraries. federal, state, and local government agencies now also rely on public libraries to provide citizens with access to and guidance in using e­government web sites, forms, and services; many government agencies simply direct citizens to the nearest public library for help. this confluence of events has created a major new social role for public libraries— guarantors of internet and e­government access. though public libraries are not the only points of free internet access in many communities, they have created the strongest commitment to providing access and help for all. by providing not only the access to technology, but also to help using the technology, libraries became internet access points, while community technology cen­ ters, which usually did not offer the same level of avail­ able assistance, failed in the late s and early s. further, as libraries not only provide internet access, but free computer access as well, they attract the people who do not own computers and do not benefit from a city’s or coffee shop’s free wi­fi. the compelling combination of free computer access, free internet access, the avail­ ability of assistance from knowledgeable librarians, the value that public librarians place on serving their local communities, and the historical trust that society places in public libraries has made libraries a critical part of the u.s. social infrastructure. without public libraries, large segments of the population would be cut off from access to the internet and e­government. while the provision of internet access for those who have no other access parallels the role of public libraries as providers of access to print materials, the matura­ tion of public libraries into internet and e­government access hubs has profound implications for the roles that public libraries are being expected to play in their communities. public libraries are trusted by their com­ munities as places that community members can turn to for unfettered information access and as places to go for information in times of need. combining this trust with the power of internet access and support makes public libraries even more critical within their local com­ munities. the trust placed in libraries is also important in balancing the lack of confidence that many citizens place in other government institutions as well as in the internet. clearly, e­government, which exists at this intersection, has its trustworthiness bolstered by the role of public libraries in its use. as patrons are able to access e­government through the library—a place that is trusted—they may have greater confidence in the gov­ ernment services they use through library computers and with the assistance of librarians. the important role of libraries in providing citizens with access to the internet, and especially to e­govern­ paul t. jaeger (pjaeger@umd.edu) is an assistant professor and director of the center for information policy and electronic government at the college of information studies of the university of maryland, college park. kenneth r. fleischmann (kfleisch@umd.edu) is an assistant professor at the college of information studies of the university of maryland, college park. paul t. jaeger and kenneth r. fleischmann public libraries, values, trust, and e-government article title | author public libraries, values, trust, and e-government | jaeger and fleischmann ment, makes natural sense given the values of the public library. these new services reflect the values traditionally upheld by public libraries, such as equal access to infor­ mation, literacy and learning, and democracy. indeed, these values likely have played a significant role in developing and sustaining public trust in public libraries as institutions. thus, to understand how public libraries have come to serve as the default site for e­government access, it is important to consider how this role builds on and reflects the public library’s enduring values. drawing upon several different strands of research, this article explores the intersections of public libraries, values, trust, and e­government. the article first exam­ ines the values of public libraries and the role that these values play in influencing popular opinion surrounding public libraries. next, the article focuses on the trust that communities place in public libraries, which builds upon the values that libraries uphold. after that, the article explores the reasons why public libraries became and remain the public access point for e­government, providing examples from the and hurricane seasons that illustrate this point in the most extreme cir­ cumstances. the article then examines the nexus of public libraries, values, trust, and e­government, further exam­ ining how the values of librarianship and the confidence that communities place in their public libraries reinforce the role of public libraries in the provision of e­govern­ ment. finally, the article explores how the e­government role of public libraries could be cultivated to improve library services through involvement in research and educational initiatives. ■ public libraries and values values can be seen as “evaluative beliefs that synthesize affective and cognitive elements to orient people to the world in which they live.” in other words, values tie together how individuals think about the world and how they feel about the world. following this definition, values are situated within individuals. although they are a result of social interaction and may be shared among individuals, values are a highly individualized and per­ sonalized phenomenon. thus, values arise at the intersec­ tion of the individual and the social, with some scholars now making a case for increasing the emphasis placed on values in the social sciences. recently, many scholars and commentators have focused on the values of librar­ ies, most notably former ala president michael gorman, who has written extensively on the topic. gorman focuses on library values in response to what he views as a disconnect between library practitioners and academics. he argues that library­science programs are becoming increasingly detached from reality, and that one way to ground library science, as well as the library profession, is through an emphasis on the values of librar­ ianship, which demonstrate the core, enduring values of the profession. he explains that values, on the one hand, should provide a foundation for interaction and mutual understanding among members of a profession; on the other hand, they should not be viewed as immutable, but rather as sufficiently flexible to match the changing times. he lists eight central values of librarianship that he views as particularly salient at present: stewardship, service, intellectual freedom, rationalism, literacy and learning, equity of access to recorded knowledge and information, privacy, and democracy. frances groen echoes gorman’s sentiments and argues that one of the major limitations of library­science programs is their lack of attention to values. she argues that library and information science (lis) programs place almost all of their educational emphasis on what librar­ ians do and how they do it, and almost none on the rea­ sons why they do what they do and why such activities are important. she identifies three fundamental library values: access to information, universal literacy, and preservation of cultural heritage, all of which she argues are also characteristics of liberal democratic societies. this argument parallels the observation that increases in information access within a society are essential to increasing the inclusiveness of the democratic process in that society. library historian toni samek focuses on another aspect of library values that is no longer as strongly emphasized—attempts to achieve neutrality in libraries. neutrality often was advocated as a cherished value, in the sense of providing equal access to all information and sources. however, samek demonstrates that libraries, on the contrary, were more likely to emphasize mainstream information sources and thus privilege them over alter­ native sources. not only has the value of neutrality been problematic in terms of how it has been implemented and mobilized in public libraries in the s and s, but it also is perhaps impossible to ever achieve in reality. the fact that neither gorman nor groen include neutrality in their listings of fundamental library values demonstrates how library values have continued to evolve as public libraries have developed as social institutions. as library values have developed, they have served to unite librarians and establish the role of public libraries in their communities. the values of librarianship have been encoded in the american library association’s (ala) library bill of rights, which strongly asserts the values of equal access and service for all patrons, nondiscrimina­ tion, diversity of viewpoint, and resistance to censorship and other abridgments of freedom of expression. the values of libraries and librarianship are one of the fac­ tors that lead communities to trust public libraries, as the following section explores. overall, further study of the information technology and libraries | december information technology and libraries | december role of values in libraries is essential, especially given the increasing role of technology in public libraries. ■ public libraries and trust exactly one half of the respondents to a pew research center study agreed with the statement “you can’t be too careful in dealing with people.” however, even in a climate where trust can be a precious commodity, public libraries are trusted by their communities. carr argues that libraries have come to earn the trust of their com­ munities because of four obligations that librarians strive to meet: to provide user­centered service, to actively engage in helping users, to connect information seekers to unexplored information sources, and to take the goal of helping users as a professional duty that is controlled first and foremost by the library user. similarly, jaeger and burnett argue that, because of its traditional defense of commonly accepted and popular values—such as free access to and exchange of information, providing a diverse range of materials and perspectives to users from across society, and opposition to government intrusions into personal reading habits—public libraries have come to be seen by members of the populace as a trusted source of information in the community. gorman argues for a direct link between the values of libraries and the trust that is instilled within them by the public, stating that one important mission for ensuring the survival of libraries and librarianship is “assuring the bond of trust between the library and the society we serve by demonstrating our stewardship and commitment, thus strengthening the mutuality of the interests of librar­ ians and the broader community.” further, a study conducted by public agenda found that “public libraries seem almost immune to the distrust that is associated with so many other institutions.” in specific terms of the internet, the public library “is a trusted community­based entity to which individuals turn for help in their online activities—even if they have comput­ ers and internet access at home or elsewhere.” in a large­ scale national survey, percent of respondents, including both users and non­users of public libraries, asserted that providing public access to the internet should be one of the highest priorities for public libraries. thus, trust in public libraries seems to carry over from other library services to provision of internet access and training. however, challenges to trust in public libraries seem to be growing in the internet age. the trusted role of pro­ tecting users’ personal information may create conflicts with the other social responsibilities of public libraries. as a result of a lack of preparedness of some librarians to deal with privacy issues, it is possible that “the trust that research shows users place in libraries is not fully repaid.” a oclc study suggests that, indeed, user trust in public libraries shows signs of weakening, as the majority of citizens place as much trust in internet search engines as they do in public libraries. further, the changes in the law following the / terror attacks that have increased the ability of the federal government to track patron activities in public libraries, such as through the usa patriot act, have raised serious concerns about privacy and freedom of expression among many public library patrons and librarians. trust in libraries also has been challenged by the impo­ sition of filters for public libraries that receive e­rate fund­ ing due to the children’s internet protection act. while internet access is no longer unfettered in libraries that have to comply with the law, public libraries have been able to prevent this law from eroding their role as trusted internet provider through ala’s vigorous legal challenge to the constitutionality of law and the rejection of e­rate funds by a large number of libraries after the supreme court upheld the constitutionality of the law. thus, the trusting rela­ tionships that public libraries have built with their com­ munities are valuable commodities that can be transferred under some circumstances from one particular service to another, yet are not inalienable rights granted to public libraries. rather, public trust is something that libraries must work hard to maintain. trust in public libraries also has served as an important cause and effect of the role of libraries in providing access to e­government. ■ public libraries and e-government public libraries are not only trusted as a means of access to the internet in general, they are trusted as a provider of access to e­government. with nearly every united states public library now connected to the internet and offer­ ing free public access, they can fill a community need of ensuring that all citizens have access to e­government and assistance using e­government services. indeed, public libraries and the internet have both improved public access to government information. this social role also is embraced by all levels of government, with government agencies often directing people with questions about their online materials to public libraries for help. as such, government agencies also trust public libraries to serve as key providers of e­ government access and training. public libraries could not have foreseen becoming the default social access point for e­government when they began to provide free public internet access in the mid­ s, due in great part to the largely separate evolution of internet access in libraries and e­government. however, they now fill this role in society, ensuring access to those who have no other means of reaching e­government and providing a safety article title | author public libraries, values, trust, and e-government | jaeger and fleischmann net of training and assistance for those who have access but need help using e­government. public libraries have developed into the social source of e­government for two reasons. the first is simply that libraries committed to the provision of public internet access in the early s and have continued to grow and improve that access so that virtually all public libraries in the united states provide free public internet access. however, presence of access alone does not account for the current role of the public library, as most public schools and government offices have internet access, and community technology centers were origi­ nally funded to create an environment that would provide computer access. a key difference in public libraries is that they are historically trusted as providers of information, including government information, to all segments of society. “the public library is one place that is culturally ingrained as a trusted source of free and open information access and exchange.” a key part of the provision of internet access in pub­ lic libraries also has been providing help. as heanue explains, “even if americans had all the hardware they needed to access every bit of government information they required, many would still need the help of skilled librarians whose job it is to be familiar with multiple systems of access to government systems.” not only is the information trusted because of the source, the help is trusted because the librarians are part of the library. as e­government has developed and the complexity has grown, this trusted help has become invaluable to many people who need to use e­government but do not feel able to on their own. in a study of both public library and internet users, the key preferences identified for public libraries included the ease of use, accuracy of informa­ tion available, and help provided by library staff. these perceptions have carried over into e­government, as the staff members not only provide help using e­government; their guidance directs users to the correct e­government sites and forms and makes using the sites an easier expe­ rience than it otherwise would be. in the era of e­government, governments internation­ ally are showing a strong preference for delivering ser­ vices via the internet, particularly as a means of boosting cost­efficiency and reducing time spent on direct interac­ tions with citizens. however, citizens show a strong preference for phone­based or in­person interactions with government representatives when they have questions or are seeking services. e­government services generally are limited by difficulties in searching for and locating the desired information, as well as lack of availability of computers and internet access to many segments of the general population. such problems are exacerbated by general lack of familiarity of the structure of government and which agencies to contact as well as many citizens’ attitudes toward technology and government. also, as e­government sites give more emphasis to presenting political agendas rather than promoting democratic par­ ticipation, users are less trusting of the sites themselves. finally, perhaps the most compelling reason for the reli­ ance on public libraries to provide access to and help with e­government is that public libraries provide support equally to all members of a community—and that free services are of most relative value to those who have the fewest resources of their own. as a result of the reliance of patrons and government agencies on the public library as a center for e­government access and assistance, public librarians have had to become de facto experts on e­government, ranging from medicare prescription plans to fema forms to immigration registra­ tion to water management registration. in one case, the involvement of a librarian who specialized in government information was necessary in a community planning pro­ cess to sort through the related e­government materials and information sources. one area where the social roles as provider of e­government and as trusted provider of information were notably intertwined was during the and hurricane seasons along the gulf coast. ■ public libraries as trusted provider of e-government public libraries have become vital access points and com­ munication hubs for many communities and, in times of emergency, are vital in helping their communities cope with the crisis. this role proved especially important in com­ munities along the gulf coast during the unprecedented and hurricane seasons, with public libraries employing their internet access to assist their communities in hurricane recovery in numerous ways. the public librar­ ies in that region described five major roles for the public library internet access in communities after a hurricane: ■ finding and communicating with dispersed and dis­ placed family members and friends; ■ completing fema forms, which are online only, and insurance claims; ■ searching for news about conditions in the areas from which they had evacuated; ■ trying to find information about the condition of their homes or places of work, including checking news sites and satellite maps; and ■ helping emergency service providers find informa­ tion and connect to the internet. the provision of e­government information and assis­ tance in filling out e­government forms was a central function of these libraries in helping their communities. the level of assistance was astounding—one mississippi library completed more than forty­five thousand fema information technology and libraries | december information technology and libraries | december applications for patrons in the first month after katrina struck—despite the fact that the libraries were not specifi­ cally prepared to offer such a service and that few library systems planned for this type of situation. furthermore, while libraries helped many communities, they could not meet the enormous needs in the affected communi­ ties. the events along the gulf coast in and revealed a serious need for the integration of local and state public entities that have large­scale coordination plans to work with the libraries. most of the functions that community organizations played in the most ravaged areas after katrina, rita, wilma, dennis, ivan, and the other major storms were completely ad hoc and unplanned. the federal gov­ ernment was of little help in the immediate aftermath of many of these situations. as such, it was the local community organizations, particularly public libraries, that used information technology (at least what was still working) to try to pick up the pieces, get aid, find the missing, and perform other vital functions. consider the following quotes from local government officials explaining the role computers and internet access in public libraries played in providing information to dev­ astated communities: our public access computers have been the only source of communicating with insurance carriers, the federal emergency management agency and other sources of aid. the greatest impact has been access to information such as fema forms and job applications that are only available via internet. this was highly visible during the aftermath of hurricanes rita & katrina. overall access to information in this rural community has been outstanding due to use of the internet. relief workers were encouraged to use the library to keep in touch with family and friends through email. . . . the library provided a fema team with local maps and help in locating areas that potentially suffered major damage from the storm. during the immediate aftermath of katrina, our com­ puters were invaluable in locating missing family, applying for fema relief (which could only be done online) and other emergency needs. for that time—the computers were a godsend. we have a large number of displaced people who are coming to rely upon the library in ways many of them never expected. i’ve had so many people tell me that they had never been to a library before they had to find someplace to file a fema application or insur­ ance claim. many of these people knew nothing about computers and would have been totally lost without the staff’s help. along with e­government access, one of the greatest affects of access to information related to searches for lost family, friends, and pets, with many libraries creating lists of individuals who had been to the library and who were being sought to help in establishing contacts between people. as one librarian stated, “our computers were invaluable in locating a missing family.” searches were conducted by patrons and by librarians helping them to locate evacuees and search for information about those who stayed behind. internet access also allowed patrons to have “contact with family members outside of the disaster area,” “communicate with family and friends,” and “stay in touch with family and friends due to lack of telephone service.” libraries used their internet access to aid rescue personnel to communicate with their agen­ cies, and even to direct emergency responders with direc­ tions, maps, and information about where people most needed help. the level of local libraries’ success in meeting the needs of their communities after the hurricanes varied widely, though. many were simply overwhelmed by the numbers of people in need and limited by the fact that they had never expected to have to act as a community lifeline in this way. the libraries that faired the best were usually in florida; they have a greater familiarity with dealing with hurricanes and thus were more prepared and had more established ties between local libraries, county governments, and state agencies. having internet access and expertise is clearly not enough. planning, coordina­ tion, experience, and government support and funding all influenced how different public libraries were able to respond after the major hurricanes. public libraries also may be able to play a role in ongoing emergency response efforts, such as the development of large­scale community response grids that coordinate citizens and emergency responders in emergencies. the greatest lesson, however, may be that public librar­ ies, as trusted providers of information technology access, particularly access to e­government, are the most local line of response in communities. the national government failed shatteringly and completely to help people after hurricane katrina, while little public libraries in and on the edges of the devastation hummed along. the local nature of the response that libraries could provide man­ aged to reach communities and members of those commu­ nities much better than national or state level responses. such local response to crises, while vital, is becoming much harder to find outside of public libraries. ■ the nexus of public libraries, values, trust, and e-government the democratically oriented core values of public librar­ ies and the trust that communities place in their public article title | author public libraries, values, trust, and e-government | jaeger and fleischmann libraries have the potential to significantly enhance and strengthen the role of public libraries in the provision of e­government. citizens who access e­government using computers in public libraries, and with the expert assistance of librarians, may have more confidence in the e­government information and services they are using as a result of their high regard for public libraries. as patrons trust that librarians will help them reach the information they need, patrons’ awareness of and confidence in e­government will increase as they learn from librarians about the types of information and services available from e­government. further, by teaching patrons what is available from and how to use e­government, librar­ ians are serving to increase the number of e­government users. because e­government is still at an early stage in its development, such positive associations could play a critical role in encouraging and facilitating its widespread acceptance and adoption. just as e­government is still in its formative stages, research on e­government also is just getting started. to date, research on e­government has focused more on technical than social aspects. for example, a meta­ analysis of peer­reviewed journal articles related to e­government revealed that the relationship between e­government and values is an important, yet to date understudied, topic. it is important to consider not only bandwidth and markup languages, but also values and trust in developing and analyzing e­government. it also is important to consider the relationship between trust in e­government and the potential for increasingly participatory democracy. trust can be seen as “centrally positioned at the nexus between the primarily internally driven administrative reforms of e­government’s architecture and the related, more exter­ nally rooted pressures for e­governance reflected in widening debates on openness and engagement.” similarly, “citizen engagement can help build and strengthen the trust relationship between governments and citizens.” through e­government, it is possible to facilitate citizen participation in government through the bidirectional interactive potential of the internet, making it possible to move toward strong democracy. greater faith in democracy can potentially significantly increase citizen trust in e­government. at the same time that we consider all of these impor­ tant issues related to e­government, it is important not to lose sight of the critical role that public libraries play in the provision of e­government. further, it is necessary to make certain that public libraries receive credit and support for the work that they do in providing access to and help with e­government. as demonstrated above, public libraries are uniquely and ideally situated to ensure access to and assistance in using e­government information and services. however, this activity is not sustainable without the recognition and resources that must accompany this role. the conclusion addresses this important point in more detail. ■ conclusions and future directions the evolution of the public library into an e­government access point has occurred without the direct intention of public libraries and without their involvement in policy decisions related to these new social roles. as with the need to become more active in encouraging the develop­ ment of technologies to help libraries fulfill these social expectations, public libraries also must become more involved in the policy­making process and in seeking financial and other support for these activities. public libraries have to demand a voice not only to better con­ vey their critical role in the provision e­government, but to help shape the direction of the policy­making process to ensure more government support for the access to and help with e­government that they provide. public libraries have taken on these responsibilities without receiving additional funding. while the provi­ sion of internet access alone is a major expense for public libraries, the reliance of government agencies on public libraries as the public support system for e­government adds very significant extra burdens to libraries. in a survey of florida public libraries, for example, . percent indicated that they receive no support from an outside agency to support the e­government services the library provides, despite the fact that . percent of responding libraries indicated that the use of e­govern­ ment in the library had increased overall library usage. this lack of outside support has resulted in public librar­ ies in different parts of the country having widely varying access to the internet. the reality is that public libraries are expected by patrons and government agencies to fulfill this social role, whether or not any support—financial, staffing, or training—is provided for this role. the vital roles that public libraries played in the aftermath of the major hur­ ricanes of the and seasons may have perma­ nently cemented the public and government perception of public libraries as hubs for e­government access. while public libraries have become the unofficial uni­ versal access point for e­government and are trusted to serve as a vital community response and recovery agency during emergencies, they do not receive funding or other forms of external assistance for these functions. public libraries need to become involved in and encourage plans and programs that will serve to sustain these essential and inextricably linked activities, while also bringing some level of financial, training, and staffing support for these roles. the tremendous efforts and successes of public librar­ ies in the aftermath of the and hurricanes has information technology and libraries | december information technology and libraries | december earned libraries a central position to e­government and emergency planning at local, state, and federal levels. in those emergency situations, public libraries were able to serve their communities in a capacity that was far beyond the traditional image of the role of libraries, but these emergency response roles are as significant as anything else libraries could do for their communities. in order to continue fulfilling these roles and adequately performing other expected functions, public libraries need to push not only for financial support, but also for a greater role in planning and decision­making related to e­government services as well as emergency response and recovery at all levels of government. if strategic plans and library activities have a consis­ tent message about the need for support, the interrelated roles of trusted source of local information, e­government access provider, and community­response information and coordination center can make a compelling argument for increases in funding, support, and social standing of public libraries. the most obvious source of further sup­ port for these activities would be the federal government. amazingly, federal government support accounts for only about percent of public library funding. given that federal government agencies are already relying on public libraries to ensure access to e­government and fos­ ter community response and recovery in times of emer­ gencies, federal support for these social roles of the public library clearly can and should be increased significantly. state libraries, cooperatives, and library networks already work to coordinate funding and activities related to certain programs, such as the e­rate program. these same library collectives may be able to work together to promote the need for additional resources and coor­ dinate those resources once they are attained. private and public partnerships offer another potential means of support for these library activities. with its strong historical and current connections to technology and libraries, the bill and melinda gates foundation might be a very important partner in funding and facilitating the increased role that public libraries play in providing access to and help with e­government. the search for additional funding to support e­government provision should not only focus on funds for access and training, but also on funds for research about how to better meet individual and community e­government needs and the affects of e­government provision by public libraries on individuals and communities. regardless of what approaches are taken to find­ ing greater support, however, public libraries must do a better job of communicating their involvement in the provision of e­government to governments and private organizations in order to increase support. such commu­ nications will need to be part of a larger strategy to define a place within public policy that gives public libraries a voice in e­government issues. if public libraries are going to fulfill this social role, they must become a greater pres­ ence in the national policy discourse surrounding e­gov­ ernment. to increase their support and standing in policy discourse, libraries must not be hesitant in reminding the public and government officials of their successes after emergencies and in providing the social infrastructure for e­filing of taxes, enrolling in medicare prescription drug plans, and myriad other routine e­government activities. in many societies, e­government has come to be seen by many citizens and governments as a force that will enhance democratic participation, more closely link citizens and their representatives, and help disadvan­ taged populations become more active participants in government and in society. e­government is seen by many as having “the potential to fundamentally change a whole array of public interactions with government.” while the e­government act of and president’s e­government management agenda have emphasized the transformative effect of e­government, thus far it has primarily been used as a way to make information available, provide forms and electronic filing, and distrib­ ute the viewpoints of government agencies. however, many citizens do look to e­government as a valuable source of information, considering e­government sites to be “objective authoritative sources.” currently, the primary reason that people use e­government is to gather information. in the united states, percent of internet users in the united states believe e­government to be the best source for government information, percent of americans expect that information they are seeking will be on a government site, and million americans seek political information online everyday. public satisfaction with the e­government services available, however, is limited. as commercial sites are developing faster and provide more innovative services than e­government sites, public satisfaction with gov­ ernment web sites is declining. public confidence in government web sites also has declined as much of the public policy related to e­government since / has been to reduce access to information through e­govern­ ment. the types of information that have been affected include many forms of socially useful information, from scientific information to public safety information to information about government activities. for these and other reasons, the majority of citizens, even those with a high­speed internet connection at home, seeking govern­ ment information and services prefer to speak to a person directly in their contacts with the government. in many cases, people turn to public librarians to serve as the per­ son involved in e­government contacts. further, when people struggle with, become frustrated by, or reject e­government services, they turn to public libraries. every year, public libraries deal with huge num­ bers of patrons needing help with online taxes, and the medicare prescription drug plan sign­up period resulted in article title | author public libraries, values, trust, and e-government | jaeger and fleischmann an influx of seniors to public libraries seeking help in using the online registration system. for example, during the tax season, virginia discontinued the distribution of free print copies of tax forms to encourage use of the online system. instead, citizens of the state flooded public librar­ ies, assuming that libraries could find them print copies of the forms, which of course the libraries did. it seems unlikely, however, that the same government officials pushing the use of e­government are aware of the roles of public libraries in helping citizens with day­to­day e­government use. further, the enormous social roles of public libraries in emergency response in communities, such as during the and hurricane seasons, are far from widely known among government officials. to encourage the provision of external funding, the develop­ ment of targeted support technologies, and policy sup­ port for these social roles, public libraries must make the government and the public better aware of these roles and what is needed to ensure that the roles can be fulfilled. similarly, there is an extremely important role for lis programs in ensuring public libraries can meet community expectations for e­government provision. lis program graduates need to be prepared to help patrons access and use e­government information and services. as govern­ ment activities move primarily or exclusively online, patrons will increasingly seek help with e­government from public libraries. lis programs must ensure that grad­ uates are ready to serve patrons in this capacity. in , the college of information studies at the university of maryland became the first ala­accredited school to offer a concentration in e­government as part of the master of library science program. the goal of this concentration is to prepare future librarians who wish to specialize in e­government, which will be an area of increasing and sig­ nificant need as more government information and services move online and more government agencies rely on public libraries to ensure access to e­government. lis programs need to prioritize finding ways to incorporate the teaching of issues related to e­government in public libraries as new concentrations or courses, or into existing courses. the provision of e­government is an important role of public libraries that is likely to increase significantly, and gradu­ ates of lis programs need to be prepared to meet patrons’ e­government information needs. further, lis faculties also can support public libraries in their e­government access and training roles by focusing more research on the intersections of public libraries and e­government. ultimately, the role of the trusted and valued public provider of e­government access creates many financial and staffing obligations and social responsibilities, but it also is a tremendous opportunity for public libraries. fighting against censorship efforts in the s estab­ lished the public perception of libraries as guardians of the first amendment during the mccarthy era. working to ensure access and the ability to use e­government is creating new public perceptions of libraries as guardians of equal access in new but just as socially meaningful ways. rather than needing to ponder whether the emer­ gence of the internet will limit or remove the relevance of public libraries, the advent of e­government has created a brand new and very significant role that public libraries can play in serving their communities. given the empha­ sis that governments are placing on moving information and services online, patrons will continue to need access to and assistance in using e­government. the trust and values that have long been associated with public libraries are evolving to include the social expectations of the provision of access to and training for e­government by public libraries. in the same ways that patrons have learned to trust public libraries to provide equal access to print information sources, they now have learned to trust that libraries can provide equal access to e­government information. it seems that citizens will regu­ larly be turning to public libraries for help with mundane e­government activities, such as finding forms and filing taxes, as well as with the most pressing e­government activities, as was demonstrated in the aftermath of hur­ ricanes katrina and rita. because the trust in and values of public libraries have set the stage for the emerging role of libraries in e­government, public libraries need to work to ensure the availability of the support, education, and policy decisions that they need to serve their communities in this new and vital role in situations ranging from every­ day information needs to the most extreme circumstances. in spite of the costs associated with serving as the public’s e­government access center, acting as the social guarantor of equal access to e­government emphatically demonstrates that public libraries will continue to be a central part of the infrastructure of society in the internet age. public libraries now must learn to articulate better the social roles they are playing and the types of support they need from lis programs, funding agencies, and gov­ ernment agencies to continue playing these roles. ■ acknowledgment the authors of this paper have worked with several col­ leagues on projects related to the ideas discussed in this paper. the authors would particularly like to thank john carlo bertot, lesley a. langa, charles r. mcclure, jennifer preece, yan qu, ben shneiderman, and philip fei wu. references and notes . margaret mooney marini, “social values and norms,” encyclopedia of sociology, edgar f. borgatta and marie l. borgatta, eds., (new york: macmillan, ). information technology and libraries | december information technology and libraries | december . steven hitlin and jane allyn piliavin, “values: reviv­ ing a dormant concept,” annual review of sociology ( ): – . . michael gorman, our singular strengths: meditations for librarians (chicago: ala, ); michael gorman, our enduring values: librarianship in the st century (chicago: ala, ); michael gorman, our own selves: more meditations for librarians (chicago: ala, ). . gorman, our enduring values. . frances k. groen, access to medical knowledge: librar- ies, digitization, and the public good (lanham, md.: scarecrow, ). . elizabeth smith, “equal information access and the evo­ lution of american democracy,” journal of educational media and library sciences , no. ( ): – . . toni samek, intellectual freedom and social responsibility in american librarianship, – (jefferson, n.c.: mcfarland, ). . pam scott, evelleen richards, and brian martin, “cap­ tives of controversy: the myth of the neutral social researcher in contemporary scientific controversies,” science, technology, and human values ( ): – . . american library association, “library bill of rights,” www.ala.org/ala/oif/statementspols/statementsif/librarybill­ rights.htm (accessed may , ). . kenneth r. fleischmann, “digital libraries with embed­ ded values: combining insights from lis and science and technology studies,” library quarterly (in press); kenneth r. fleischmann, “digital libraries and human values: human­ computer interaction meets social informatics,” proceedings of the th annual conference of the american society for infor­ mation science and technology, milwaukee, wisc., . . pew research center, americans and social trust: who, where, and why (washington, d.c.: pew research center, ), http://pewresearch.org/assets/social/pdf/socialtrust.pdf, . . david wildon carr, “an ethos of trust in information service,” in ethics and electronic information: a festschrift for stephen almagno, barbara rockenbach and tom mendina, eds., – (jefferson, n.c.: mcfarland, ). . paul t. jaeger and gary burnett, “information access and exchange among small worlds in a democratic society: the role of policy in redefining information behavior in the post­ / united states,” library quarterly , no. ( ): – . . gorman, our enduring values, . . public agenda, long overdue: a fresh look at public and leadership attitudes about libraries in the st century (new york: public agenda, ), , www.publicagenda.org/research/ pdfs/long_overdue.pdf (accessed may , ). . john carlo bertot et al., “public access computing and internet access in public libraries: the role of public librar­ ies in e­government and emergency situations,” first monday , no. ( ), www.firstmonday.org/issues/issue _ /bertot (accessed may , ). . public agenda, long overdue. . nancy zimmerman and feili tu, “it is not just a matter of ethics ii: an examination of issues related to the ethical provi­ sion of consumer health services in public libraries,” ethics and electronic information: a festschrift for stephen almagno, barbara rockenbach and tom mendina, eds., – (jefferson, n.c.: mcfarland, ). . paul sturges and ursula iliffe, “preserving a secret garden for the mind: the ethics of user privacy in the digital library,” ethics and electronic information: a festschrift for stephen almagno, barbara rockenbach and tom mendina, eds., – (jefferson, n.c.: mcfarland, ), . . online computer library center, inc. (oclc), perceptions of libraries and information resources: a report to the oclc mem- bership (dublin, ohio: oclc, ). . jaeger and burnett, “information access and exchange among small worlds in a democratic society”; paul t. jaeger et al., “the usa patriot act, the foreign intelligence surveil­ lance act, and information policy research in libraries: issues, impacts, and questions for library researchers,” library quar- terly , no. ( ): – . . children’s internet protection act, public law – . . paul t. jaeger, john carlo bertot, and charles r. mcclure, “the effects of the children’s internet protection act (cipa) in public libraries and its implications for research: a statistical, policy, and legal analysis,” journal of the american society for information science and technology , no. ( ): – ; paul t. jaeger et al., “cipa: decisions, implementation, and impacts,” public libraries , no. ( ): – . . bertot et al., “public access computing and internet access in public libraries”; john carlo bertot et al., “drafted: i want you to deliver e­government,” library journal , no. ( ): – ; john carlo bertot et al., public libraries and the internet : study results and findings (tallahassee, fla.: infor­ mation institute, ), www.ii.fsu.edu/plinternet_reports.cfm (accessed may , ). . nancy kranich, “libraries, the internet, and democracy,” libraries & democracy: the cornerstones of liberty, nancy kranich, ed., – (chicago: ala, ). . bertot et al., “public access computing and internet access in public libraries”; bertot et al., “drafted.” . bertot et al., public libraries and the internet . . jaeger and burnett, “information access and exchange among small worlds in a democratic society,” . . anne heanue, “in support of democracy: the library role in public access to government,” information, libraries, and democracy: the cornerstones of liberty, nancy kranich, ed. (chi­ cago: ala, ), . . george d’elia et al., “the impact of the internet on public library uses: an analysis of the current consumer market for library and internet services,” journal of the american society for information science and technology , no. ( ): – ; eleanor jo rodger, george d’elia, and corrine jorgensen, “the public library and the internet: is peaceful coexistence pos­ sible?,” american libraries , no. ( ): – . . w. e. ebbers, w. j. pieterson, and h. n. noordman, “elec­ tronic government: rethinking channel management strate­ gies,” government information quarterly (in press). . ibid. . awdhesh k. singh and rajendra sahu, “integrating inter­ net, telephones, and call centers for delivering better quality e­governance to all citizens,” government information quarterly (in press). . paul t. jaeger and kim m. thompson, “e­government around the world: lessons, challenges, and new directions,” government information quarterly , no. ( ): – ; paul t. jaeger and kim m. thompson, “social information behavior article title | author public libraries, values, trust, and e-government | jaeger and fleischmann and the democratic process: information poverty, normative behavior, and electronic government in the united states,” library & information science research , no. ( ): – . . paul t. jaeger, “deliberative democracy and the con­ ceptual foundations of electronic government,” government information quarterly , no. ( ): – ; paul t. jaeger, “information policy, information access, and democratic partic­ ipation: the national and international implications of the bush administration’s information politics,” government information quarterly (in press). . bertot et al., “public access computing and internet access in public libraries”; bertot et al., “drafted.” . aimee c. quinn and laxmi ramasubramanian, “infor­ mation technologies and civic engagement: perspectives from librarianship and planning,” government information quarterly (in press). . bertot et al., public libraries and the internet ; paul t. jaeger et al., “the and gulf coast hurricanes: evolv­ ing roles and lessons learned for public libraries in disaster preparedness and community services,” public library quarterly (in press). . bertot et al., “drafted.” . jaeger et al., “the and gulf coast hurricanes.” . ibid. . ibid. . michael arnone, “storm watch : ready or not,” fed- eral computer week, june , , www.fcw.com/print/ _ / news/ ­ .html (accessed may , ). . jaeger et al., “the and gulf coast hurricanes.” . bertot et al., “public access computing and internet access in public libraries.” . jaeger et al., “the and gulf coast hurricanes.” . ibid. . ibid. . bertot et al., “public access computing and internet access in public libraries.” . paul t. jaeger et al., “ .gov: harnessing e­government, mobile communication technologies, and social networks to promote community participation in emergency response,” telecommunications policy (in press); ben shneiderman and jenny preece, “ .gov: community response grids,” science ( ): . . kim viborg andersen and helle zinner henriksen, “e­government research: capabilities, interaction, orientation, and values,” current issues and trends in e-government research, donald f. norris, ed., – (hershey, pa.: cybertech, ). . jeffrey roy, “e­government in canada: transition or trans­ formation?” current issues and trends in e-government research, donald f. norris, ed., – (hershey, pa.: cybertech, ), . . oecd e­government studies, the e-government imperative (danvers, mass.: organization for economic co­operation and development, ), . . bruce barber, strong democracy (berkeley, calif.: univ. of california pr., ). . bertot et al., public libraries and the internet . . charles r. mcclure et al., e-government and public librar- ies: current status, meeting report, findings, and next steps (tallahassee, fla.: information use management and policy institute, ), www.ii.fsu.edu/announcements/e­gov / egov_report.pdf (accessed may , ). . paul t. jaeger et al., “public libraries and internet access across the united states: a comparison by state from to ,” information technology and libraries , no. ( ): – . . jaeger et al., “the and gulf coast hurricanes.” . bertot et al., “drafted.” . jaeger et al., “public libraries and internet access across the united states.” . beth simone noveck, “designing deliberative democracy in cyberspace: the role of the cyber­lawyer,” boston university journal of science and technology ( ): – . . s. h. holden and l. i. millett, “authentication, privacy, and the federal e­government,” information society ( ): . . e­government act of , p.l. – ; jaeger, “delibera­ tive democracy and the conceptual foundations of electronic government”; e-government strategy: implementing the president’s management agenda for e-government (washington, d.c.: egov, ), www.whitehouse.gov/omb/egov/ egov_strat.pdf (accessed may , ). . anderson office of government services, a usability analysis of selected federal government web sites (anderson office of government services: washington, d.c., ), . . christopher g. reddick, “citizen interaction with e­gov­ ernment: from the streets to servers?,” government information quarterly , no. ( ): – . . john b. horrigan, politics online (washington, d.c., pew internet & american life project, ); john b. horrigan and lee rainie, counting on the internet (washington, d.c., pew internet & american life project, ). . stephen barr, “public less satisfied with government websites,” washington post, mar. , , www.washingtonpost. com/wp­dyn/content/article/ / / /ar . html (accessed may , ). . lotte e. feinberg, “foia, federal information policy, and information availability in a post­ / world,” govern- ment information quarterly ( ): – ; elaine l. halchin, “electronic government: government capability or terrorist resource,” government information quarterly ( ): – : harold c. relyea and elaine l. halchin, “homeland security and information management,” the bowker annual: library and trade almanac , d. bogart, ed., – (medford, n.j.: infor­ mation today, ). . jaeger, “information policy, information access, and democratic participation.” . john b. horrigan, how americans get in touch with govern- ment (washington, d.c., pew internet & american life project, ). . bertot et al., “public access computing and internet access in public libraries”; bertot et al., “drafted.” . the description of the university of maryland’s e­gov­ ernment master’s program is available at www.clis.umd.edu/ programs/egov.shtml. . jaeger and burnett, “information access and exchange among small worlds in a democratic society.” information technology and libraries | december author id box for column layout column titlecommunications afghanistan digital library initiative: revitalizing an integrated library system yan han and atifa rawan this paper describes an afghanistan digital library initiative of building an integrated library system (ils) for afghanistan uni- versities and colleges based on open-source software. as one of the goals of the afghan equality digital libraries alliance, the authors applied systems analysis approach, evaluated different open-source ilss, and customized the selected software to accom- modate users’ needs. improvements include arabic and persian language support, user interface changes, call number label print- ing, and isbn- support. to our knowl- edge, this ils is the first at a large academic library running on open-source software. the last quarter­century has been devastating for afghanistan, with an uninterrupted period of inva­ sions, civil wars, and oppressive regimes. “since , the education system was virtually destroyed on all levels. schools and colleges were closed, looted, or physically reduced; student bodies and faculties were emptied by war, migration, and eco­ nomic hardship; and libraries were gutted.” kabul university (ku), for example, was largely demolished by and completely closed down in . it is universally recognized that afghanistan desperately needs trained faculty, teachers, librarians, and staff. the current state of the higher education system is one of dramatic destruction and deteriora­ tion. based on rawan’s assessments of ku library, most of its collections were damaged or destroyed. she found that there were approximately , to , books in english, , to , books in persian, and , theses in persian. none of these collections have manual or online catalog records. the library has eigh­ teen staff members, but not all are fully trained in library activities. rebuilding the educational infra­ structure in afghanistan is essential. afghan equality digital libraries alliance the university of arizona (ua) library has been involved in rebuilding academic libraries in afghanistan since april . in , we were invited to be part of the digital libraries alliance (dla) as part of the afghan equality alliances: st century universities for afghanistan initiative funded by the usaid and washington state university. dla’s goal is to build the capacity of afghan libraries and librarians to work with open source digital libraries platforms; and to provide and enhance access to schol­ arly information resources and open content that all afghanistan univer­ sities can share. revitalizing the afghan ils an integrated library system (ils) usually includes several critical com­ ponents, such as acquisitions, cat­ aloging, catalog (search and find), circulation, and patron management. traditionally it has been the center of any library. recent developments in digital libraries have resulted in dis­ tributed systems in libraries, and the ils is treated as one of many digital library systems. it still is critical to have a centralized ils to provide a primary way to access library­owned materials for afghanistan universi­ ties and colleges. other services, such as interlibrary loan and other digital library systems, can be further devel­ oped to extend libraries’ services to users and communities. the ua library is working collab­ oratively with other dla members, including universities around the world and universities in afghanistan. one of the goals is to develop a digital library environment, includ­ ing a centralized ils for four aca­ demic universities in kabul (kabul university, polytechnic university, kabul medical university, and kabul education university). in the future, the ils will include other regional institutions throughout afghanistan. the ils will support , students and , faculty in afghan universi­ ties and colleges. overview of the ils market currently the ils market is primar­ ily dominated by commercial sys­ tems, such as innovative interface, endeavor, and sirsi. compared with other computing areas, open­source systems in ils are immature and limited, as there are only a few prod­ ucts available, and most of them do not have the full features of an ils. however, they are providing a valu­ able alternative to those costly com­ mercial systems. based on the availability of exist­ ing funding, experiences with com­ mercial vendors, and consideration of vendor supports and future direc­ tions, the authors decided to build a digital library infrastructure with the open concept (open access, open source, and open standards). the decision is widely influenced by glo­ balization, open access, open source, open standards, and increasing user expectations. at the same time, the decision gives us an opportunity to develop and integrate new tools and services for libraries as suggested by the university of california. koha is probably the most renowned open­source ils. it is yan han (hany@u.library.arizona.edu) is systems librarian and atifa rawan (rawana@u.library.arizona.edu) is librarian at the university of arizona libraries, tucson. afghanistan digital library initiative | han and rawan a full­featured ils, developed in new zealand and first deployed in horowhenua library trust in . so far koha has been running in a few public and special libraries. the underlying architecture is the linux, apache, mysql, and perl (lamp) stack. building on a simi­ lar lamp (linux, apache, mysql, and php) architecture, openbiblio has a relatively short history, releas­ ing its first beta . . version in and currently in beta . . version. webils is an open­source ils based on unesco’s cds/isis database, developed by the institute for computer and information engineering in poland. the software has some ils features, including cataloging, catalog (search and find), loan, and report modules. weblis must run on windows and window­ based web servers, such as xitami/ microsoft iis and isis database. gnuteca, another open­source ils widely deployed in south america universities, was developed in brazil. as with webils, it has some ils features, such as cataloging, cata­ log, and loan; however, the software interface is written in portuguese, which presents a language barrier for u.s. and afghanistan users. the paper open source integrated library systems provides a good overview of other systems. systems analysis the authors adopted systems analy­ sis by taking account of afghan col­ lections, users’ needs, and systems functionality required to perform essential library operations. koha was chosen as the base software, due to its functionality, maturity, and support. some of the reasons are: ■ the software architecture is open­ source lamp, which is popular, stable, and predominant. ■ our staff have skills in these open software systems. ■ it is a full­featured open­source ils. certain components, such as multiple branch support and user management, are critical. ■ two large public libraries serv­ ing population of , users in new zealand and united states have been running their ils on koha for a few years. the soft­ ware is stable, and most bugs have been fixed. ■ koha has a mailing list that is used by koha developers and users as a communication tool to ask and answer questions. kabul universities have com­ puter science faculty and students who have the capacity to participate in the development. due to working schedules and locations, we prefer to develop and maintain the system in the ua library. the technical project team consists of three people: yan han, who is responsible for manag­ ing the overall implementation and development in the open source ils system; one part­time (twenty hours per week) student developer whose major task is to develop and man­ age source code; and a temporary student (ten hours per week for two months) responsible for translating english to farsi and dari. testing tasks, such as unit testing and sys­ tem testing, are shared by all mem­ bers of the team. major challenges farsi and dari languages support koha version . cannot correctly handle east asian language records, including farsi and dari records. supporting persian, farsi, and dari records is a very important require­ ment, as these afghan universities have quite a few persian and dari materials. koha generates a web­ based graphical user interface (gui) through perl included templates that use a html meta tag with western character set (iso­ ­ ) to encode characters. browsers such as internet explorer and firefox use the meta tag to decode characters with a predefined character set. therefore, other characters, such as arabic and persian as well as chinese would not be displayed correctly. the perl tem­ plates were identified and modified to allow characters to be encoded in unicode, and this solved the prob­ lem. persian and dari characters can be entered into the cataloging module and displayed correctly in the gui. however, we should understand the limitations of this approach when dealing with other east asian character sets, such as chinese characters. only frequently used characters can be represented. a project of academia sinica is one of the efforts to deal with , unique chinese characters. farsi/dari gui as the project is designed for local afghanistan users, there is a need for a farsi and dari gui. the current version of koha does not have such an interface, and we decided to create a new farsi/dari gui for the opac. the koha system’s internal structure is logically arranged; therefore, our development work in translation is not difficult to manage. the transla­ tion student translates english words in perl template files into farsi and dari. at the same time he works with the developer to make sure it is dis­ played correctly in the opac. figure is the screenshot of the gui. other improvements we further developed a spine label printing module and integrated the module into the ils, as there is no such function provided. the module allows library staff to print one or more standardized labels ( . inches high by inch wide) with oclc formats on gaylord lsl paper, which has fifty­six labels per sheet. information technology and libraries | december lstaff can select an appropriate label slot to start and print out his or her choices of labels through the web preview feature. this feature eases library staff operations and provides cost savings for label papers. isbn­ replaced isbn­ after january , , and any ils has to be able to handle the new isbn­ . our ils has been improved to han­ dle both isbn standards. thanks to koha’s delegation of the gui and major functionality, interfaces such as fonts and web pages can be modi­ fied through the templates and css. a z . service has been configured to allow users to search other librar­ ies’ catalogs. hardware and software support afghanistan is still developing its fun­ damental infrastructure: electricity, transportation, and communication. when considering buying hardware for the ils, difficult issues, such as server services and computer parts, have to be solved. even international it companies, such as dell, hp, and ibm, have very limited services and support in afghanistan. regarding software and system support, our strategies are to: ■ maintain and develop the open source software at the ua library by the project team; ■ run one server in kabul, afghanistan, administrated by a local system administrator. ■ run one server in the ua library administrated by the library’s system administrator. cost we estimated our overall cost for building the open­source system is low and reasonable. the system is currently run­ ning on a dell server ($ , for ghz cpu, gb ram, and five gb hard drives), kernel built debian linux (free), apache (free), mysql (free), and perl (free). han spends four hours per week for coor­ dination, communication, and man­ agement of the project. the student developer works twenty hours per week for development and mainte­ nance, while the translation student will spend one hundred hours for translation. conclusion revitalizing an afghan ils is the first important goal to build digital library initiatives for the afghanistan higher education system. by under­ standing afghan university librar­ ies, collections, and users, the ua library is working with other dla members to build the open source ils. the new farsi and dari user interface, language support, and other improvements have been made to meet needs of afghan uni­ versities and colleges. the cost of using and developing existing open source software is reasonable. acknowledgments we thank usaid, washington state university, and other dla mem­ bers for providing support. this work was supported by usaid and washington state university. references and notes . nazif sharani et. al., conference transcription, conference on strate­ gic planning of higher education for afghanistan, , indiana university, bloomington, oct. – . . atifa rawan, transformation in afghanistan: rebuilding libraries, paper presented at azla conference, mesa, ariz., oct. – , . . the university of california libraries, rethinking how we provide bibliographic services for the university of california, , http://libraries.univer­ sityofcalifornia.edu/sopag/bstf/final. pdf. . eric anctil and jamshid beheshti, open source integrated library systems: an overview, , www.anctil.org/users/ eric/oss ils.html (accessed nov. , ). . derming juang et al., “resolving the unencoded character problem for chinese digital libraries,” proceedings of the th acm/ieee-cs joint conference on digital libraries, jcdl , denver (june – , ): – (new york: acm pr., ). figure : afghanistan academic libraries union catalog in farsi/dari lita cover , cover , cover index to advertisers information technology and libraries | september w elcome to my first ital president’s column. each president only gets a year to do these col­ umns, so expectations must be low all around. my hope is to stimulate some thinking and conversation that results in lita members’ ideas being exchanged and to create real opportunities to implement those ideas. my first column i thought i would keep short and sweet, and discuss just a few of the ideas that have been rattling around in my head since the midwinter lita town meeting, which have been enhanced by a number of discussions among librarians over the last six months. with any luck, these thoughts might have some bearing on what any of those ideas could mean to our organization. first off, i don’t think i can express how weird this whole presidential appellation is to me. i am extremely proud to be associated with lita, and honored and surprised at being elected. i come from a consortia envi­ ronment and an extremely flat organization. solving problems is often a matter of throwing all the parties in a room together and hashing it out until solutions are arrived at. i’ve been a training librarian for quite a while now, and pragmatic approaches to problem solving are my central focus. i’m a consortia wrangler, a trainer, and a technology pusher, and i hope my approach is, and will be, to listen hard and then see what can be accomplished. so in my own way, i find being president kind of on the embarrassing side. it’s like not knowing what to do with your hands when you’re speaking in public. at the lita town meeting (http://litablog .org/ / / /lita­town­meeting­ ­report/) it was pretty obvious that members want community in all its various forms, face­to­face in multiple venues and online in multiple venues. it’s also pretty obvious from the studies done by pew internet and american life and by oclc that our users, and in particular our younger users, really want community. the web . and the library . movements are responses to that desire. as a somewhat flippant observation, we spent a generation educating our kids to work in groups, and now we shouldn’t be sur­ prised that they want to work and play in groups. many of us work effectively in collaborative groups everyday. we find it exciting, productive, and even fun. it’s an environment that we would like to create for our patrons, in­house and virtually. it’s what we would like to see in our association. having been to every single top tech trends program and listened to the lita trendsters, one theme that often comes up is that complaining about the systems our ven­ dors deliver can at times be pointless, because they sim­ ply deliver what we ask for. there is of course a corollary to this. once a system is in the marketplace, adding func­ tionality often becomes centered around the low­hang­ ing fruit. as a fictitious example, a vendor might easily add the ability to change the colors of the display to the patron, but adding a shelf list browse might take serious coding to create. so through discussions and rfp, we ask for and get the pretty colors while the browsing function waits, a form of procrastination. so then does innovation come only when all the low­hanging fruit has finally been plucked, and there’s nothing else to procrastinate on? as social organizations, libraries, ala, lita and other groups, it appears that we have plucked all the low­hanging fruit of web . . e­mail and static web pages have been done to death. as a pragmatist, what concerns me most is implementation. what delivery systems should and can we adopt and develop to fulfill the promise of services we’d like? can we ensure that barriers to participation are either eliminated or so low as to include everyone? i like to think that web . is innovation toward mirroring how we personally want to work and play and how we want our social structures to perform. so how can we make lita mirror how we want to work and play? i do know it’s not just making everything a wiki. mark beatty (mbeatty@wils.wisc.edu) is lita president / and trainer, wisconsin library services, madison. president’s column mark beatty testing information literacy in digital environments | katz despite coming of age with the internet and other tech- nology, many college students lack the information and communication technology (ict) literacy skills neces- sary to navigate, evaluate, and use the overabundance of information available today. this paper describes the development and early administrations of ets’s iskills assessment, an internet-based assessment of informa- tion literacy skills that arise in the context of technology. from the earliest stages to the present, the library com- munity has been directly involved in the design, develop- ment, review, field trials, and administration to ensure the assessment and scores are valid, reliable, authentic, and useful. t echnology is the portal through which we interact with information, but there is growing belief that people’s ability to handle information—to solve problems and think critically about information—tells us more about their future success than does their knowledge of specific hardware or software. these skills—known as information and communications technology (ict) literacy—comprise a twenty­first­century form of literacy in which researching and communicating information via digital environments are as important as reading and writing were in earlier centuries (partnership for st century skills ). although today’s knowledge society challenges stu­ dents with overabundant information of often dubious quality, higher education has recognized that the solution cannot be limited to improving technology instruction. instead, there is an increasingly urgent need for students to have stronger information literacy skills—to “be able to recognize when information is needed and have the ability to locate, evaluate, and use effectively the needed information” (american library association )—and apply those skills in the context of technology. regional accreditation agencies have integrated information lit­ eracy into their standards and requirements (for example, middle states commission on higher education ; western association of schools and colleges ), and several colleges have begun campuswide initiatives to improve the information literacy of their students (for example, the california state university ; university of central florida ). however, a key challenge to designing and implementing effective information lit­ eracy instruction is the development of reliable and valid assessments. without effective assessment, it is difficult to know if instructional programs are paying off—whether students’ information literacy skills are improving. ict literacy skills are an issue of national and inter­ national concern as well. in january , educational testing service (ets) convened an international ict literacy panel to study the growing importance of exist­ ing and emerging information and communication tech­ nologies and their relationship to literacy. the results of the panel’s deliberations over fifteen months highlighted the growing importance of ict literacy in academia, the workplace, and society. the panel called for assessments that will make it possible to determine to what extent young adults have obtained the combination of techni­ cal and cognitive skills needed to be productive mem­ bers of an information­rich, technology­based society (international ict literacy panel ). this article describes ets’s iskills assessment (for­ merly “ict literacy assessment”), an internet­based assessment of information literacy skills that arise in the context of technology. from the earliest stages to the pres­ ent, the library community has been directly involved in the design, development, review, field trials, and admin­ istration to ensure the assessment and scores are valid, reliable, authentic, and useful. ■ motivated by the library community although the results of the international ict literacy panel provided recommendations and a framework for an assessment, the inspiration for the current iskills assessment came more directly from the higher educa­ tion and library community. for many years, faculty and administrators at the california state university (csu) had been investigating issues of information literacy on their campuses. as part of their systemwide information competence initiative that began in , researchers at csu undertook a massive ethnographic study to observe students’ research skills. the results suggested a great many shortcomings in students’ infor­ mation literacy skills, which confirmed librarian and classroom faculty anecdotal reports. however, clearly such a massive data collection and analysis effort would be unfeasible for documenting the information literacy skills of students throughout the csu system (dunn ). gordon smith and the late ilene rockman, both of the csu chancellor ’s office, discussed with ets the idea of developing an assessment of ict literacy that could support csu’s information competence initiative as well as similar initiatives throughout the higher edu­ cation community. irvin r. katz irvin r. katz (ikatz@ets.org) is senior research scientist in the research and development division at educational testing service. testing information literacy in digital environments: ets’s iskills assessment � information technology and libraries | september � information technology and libraries | september ■ national higher education ict literacy initiative in august , ets established the national higher education ict literacy initiative, a consortium of seven colleges and universities that recognized the need for an ict literacy assessment targeted at higher educa­ tion. representatives of these institutions collaborated with ets staff to design and develop the iskills assessment. the consortium built upon the work of the international panel to explicate the nature of ict literacy in higher education. over the ensuing months, repre­ sentatives of consortium institutions served as subject­ matter experts for the assessment design and scoring implementation. the development of the assessment followed a process known as evidence­centered design (mislevy, steinberg, and almond ), a systematic approach to the design of assessments that focuses on the evidence (student performance and products) of proficiencies as the basis for constructing assessment tasks. through the evidence­ centered design process, ets staff (psychometricians, cognitive psychologists, and test developers) and sub­ ject­matter experts (librarians and faculty) designed the assessment by considering first the purpose of the assess­ ment and by defining the construct—the knowledge and skills to be assessed. these decisions drove discussions of the types of behaviors, or performance indicators, to serve as evidence of student proficiency. finally, simulation­ based tasks designed around authentic scenarios were crafted to elicit from students the critical performance indicators. katz et al. ( ) and brasley ( ) provide a detailed account of this design and development process, illustrating the critical role played by librarians and other faculty from higher education. ■ ict literacy = information literacy + digital environments consortium members agreed with the conclusions of the international ict literacy panel that ict literacy must be defined as more than technology literacy. college students who grew up with the internet (the “net generation”) might be impressively technologically literate, more accepting of new technology, and more technically facile than their parents and instructors (oblinger and oblinger ). however, anecdotally and in small­scale studies, there is increasing evidence that students do not use technology effectively when they conduct research or communicate (rockman ). many educators believe that students today are less information savvy than earlier generations despite having powerful information tools at their disposal (breivik ). ict literacy must bridge the ideas of information literacy and technology literacy. to do so, ict literacy draws out the technology­related components of infor­ mation literacy as specified in the often­cited standards of the association of college and research libraries (acrl) (american library association ), focusing on how students locate, organize, and communicate information within digital environments (katz ). this conflu­ ence of information and technology directly reflects the “new illiteracy” concerns of educators: students quickly adopt new technology, but do not similarly acquire skills for being critical consumers and ethical producers of information (rockman ). students need training and practice in ict literacy skills, whether through general education or within discipline coursework (rockman ). the definition of ict literacy adopted by the con­ sortium members reflects this view of ict literacy as information literacy needed to function in a technological society: ict literacy is the ability to appropriately use digital technology, communication tools, and/or networks to solve information problems in order to function in an information society. this includes having the ability to use technology as a tool to research, organize, and communicate information and having a fundamental understanding of the ethical/legal issues surrounding accessing and using information (katz et al. , ). consortium members further refined this defini­ tion, identifying seven performance areas (see figure ). these areas mirror the acrl standards and other related standards, but focus on elements that were judged most central to being sufficiently information literate to meet the challenges posed by technology. ■ ets’s iskills assessment ets’s iskills assessment is an internet­delivered assess­ ment that measures students’ abilities to research, orga­ nize, and communicate information using technology. the assessment focuses on the cognitive problem­solving and critical­thinking skills associated with using technol­ ogy to handle information. as such, scoring algorithms target cognitive decision­making rather than technical competencies. the assessment measures ict literacy through the seven performance areas identified by con­ sortium members, which represent important problem­ solving and critical­thinking aspects of ict literacy skill (see figure ). assessment administration takes approx­ imately seventy­five minutes, divided into two sec­ tions lasting thirty­five and forty minutes, respectively. article title | author testing information literacy in digital environments | katz figure . components of ict literacy define: understand and articulate the scope of an information problem in order to facilitate the electronic search for information, such as by: ■ distinguishing a clear, concise, and topical research question from poorly framed questions, such as ones that are overly broad or do not otherwise fulfill the information need; ■ asking questions of a “professor” that help disambiguate a vague research assignment; and ■ conducting effective preliminary information searches to help frame a research statement. access: collect and/or retrieve information in digital environments. information sources might be web pages, databases, discussion groups, e-mail, or online descriptions of print media. tasks include: ■ generating and combining search terms (keywords) to satisfy the requirements of a particular research task; ■ efficiently browsing one or more resources to locate pertinent information; and ■ deciding what types of resources might yield the most useful information for a particular need. evaluate: judge whether information satisfies an information problem by determining authority, bias, timeliness, relevance, and other aspects of materials. tasks include: ■ judging the relative usefulness of provided web pages and online journal articles; ■ evaluating whether a database contains appropriately current and pertinent information; and ■ deciding the extent to which a collection of resources sufficiently covers a research area. manage: organize information to help you or others find it later, such as by: ■ categorizing e-mails into appropriate folders based on a critical view of the e-mails’ contents; ■ arranging personnel information into an organizational chart; and ■ sorting files, e-mails, or database returns to clarify clusters of related information. integrate: interpret and represent information, such as by using digital tools to synthesize, summarize, compare, and contrast information from multiple sources while: ■ comparing advertisements, e-mails, or web sites from competing vendors by summarizing information into a table; ■ summarizing and synthesizing information from a variety of types of sources according to specific criteria in order to compare information and make a decision; and ■ re-representing results from an academic or sports tournament into a spreadsheet to clarify standings and decide the need for playoffs. create: adapt, apply, design, or construct information in digital environments, such as by: ■ editing and formatting a document according to a set of editorial specifications; ■ creating a presentation slide to support a position on a controversial topic; and ■ creating a data display to clarify the relationship between academic and economic variables. communicate: disseminate information tailored to a particular audience in an effective digital format, such as by: ■ formatting a document to make it more useful to a particular group; ■ transforming an e-mail into a succinct presentation to meet an audience’s needs; ■ selecting and organizing slides for distinct presentations to different audiences; and ■ designing a flyer to advertise to a distinct group of users. © educational testing service. all rights reserved. information technology and libraries | september information technology and libraries | september during this time, students respond to fifteen interactive, performance­based tasks. each interactive task presents a real­world scenario, such as a class or work assignment, that frames the infor­ mation problem. students solve information­handling tasks in the context of simulated software (for example, e­mail, web browser, library database) having the look and feel of typical applications. there are fourteen three­ to five­minute tasks and one fifteen­minute task. the three­ to five­minute tasks target a single perfor­ mance area, while the fifteen­minute tasks comprise more complex problem­solving scenarios that target multiple performance areas. the simpler tasks contribute to the overall reliability of the assessment, while the more com­ plex task focuses on the richer aspects of ict literacy performance. in the assessment, a student might encounter a sce­ nario that requires him or her to access information from a database using a search engine (see figure ). the results are tracked and strategies scored based on how he or she searches for information, such as key words chosen, search strategies refined, and how well the information returned meets the needs of the task. the assessment tasks each contain mechanisms to keep students from pursuing unproductive actions in the simulated environment. for example, in an internet browsing task, when the student clicks on an incorrect link, he might be told that the link is not needed for the current task. this message cues the student to try an alter­ native approach while still noting for scoring purposes that the student made a misstep. in a similar way, the student who fails to find useful (or any) journal articles in her database search might receive an instant message from a “teammate” providing her with a set of journal articles to be evaluated. these mechanisms potentially keep students from becoming frustrated (for example, via a fruitless search) while providing the opportunity for the students to demonstrate other aspects of their skills (for example, evaluation skills). the scoring for the iskills assessment is completely automated. unlike a multiple­choice question, each simu­ lation­based task provides many opportunities to collect information about a student and allows for alternative paths leading to a solution. scored responses are pro­ duced for each part of a task, and a student’s overall score on the test accumulates the individual scored responses across all assessment tasks. the assessment differs from existing measures in sev­ eral ways. as a large­scale measure, it was designed to be administered and scored across units of an institution or across institutions. as a simulation­based assessment, the tasks go beyond what is possible in multiple­choice format, providing students with the look and feel of interactive digital environments along with tasks that elicit higher­order critical­thinking and problem­solving skills. as a scenario­based assessment, students become engaged in the world of the tasks, and the task scenarios describe the types of assignments students should be see­ ing in their ict literacy instruction as well as examples of workplace and personal information problems. ■ two levels of assessments the iskills assessment is offered at two levels: core and advanced. the core level was designed to assess readi­ ness for the ict literacy demands of college. it is targeted at high school seniors and first­year college students. the advanced level was designed to assess readiness for the ict literacy challenges in transitioning to higher­level college coursework, such as moving from sophomore to junior year or transferring from a two­year to a four­year institution. the advanced level targets students in their second or third year of post­secondary study. the key difference between the core and advanced levels is in the difficulty of the assessment tasks. tasks in the core level are designed to be easier; examinees are presented with fewer options, the scenarios are more straightforward, and the reasoning needed for each step in a task is simpler. an advanced task might require an individual to infer the search terms needed from a gen­ eral description of an information need; the correspond­ ing core task would state the information need more explicitly. in a task of evaluating web sites, the core level might present a web site with many clues that it is not figure . in the iskills assessment, students demonstrate their skills at handling information through interaction with simulated software. in this example task, students develop a search query as part of a research assignment on earthquakes. © educational testing service. all rights reserved. article title | author testing information literacy in digital environments | katz authoritative (a “.com” url, unprofessional look, content that directly describes the authors as students). the cor­ responding advanced task would present fewer cues of the web site’s origin (for example, a professional look, but careful reading reveals the web site is by students). ■ score reports for individuals and institutions both levels of the assessment feature online delivery of score reports for individuals and for institutions. the individual score report is intended to help guide students in their learning of ict literacy skills, aiding identifica­ tion of students who might need additional ict literacy instruction. the report includes an overall ict literacy score, a percentile score, and individualized feedback on the student’s performance (see figure ). the percentile compares students to a reference group of students who took the test in early and who fall within the target population for the assessment level (core or advanced). as more data are collected from a greater number of institutions, these reference groups will be updated and, ideally, approach nationally representative norms. score reports are available online to students, usually within one week. high schools, colleges, and universities receive score reports that aggregate results from the test­takers at their institution. the purpose of the reports is to provide an overview of the students in comparison with a reference group. these reports are available to institutions online after at least fifty students have taken either the core or advanced level test—that is, when there are sufficient num­ bers to allow reporting of reliable scores. figure shows a graph from one type of institutional report. users have the option to specify the reference group (for example, all students, all students at a four­year institution) and the subset of test­takers to compare to that group (for exam­ ple, freshmen, students taking the test within a particular timeframe). a second report summarizes the performance feedback of the individual reports, providing percentages of students who received the highest score on each aspect of performance (each of the fourteen short tasks are scored on two or three different elements). finally, institutions can conduct their own analyses by downloading the data of their test­takers, which include each student’s responses to the background questions, iskills score, and responses to institution­specified questions. ■ testing the test a variety of specialists contributed to the development of ets’s iskills assessment: librarians, classroom fac­ ulty, education administrators, assessment specialists, researchers, user­interface and graphic designers, and systems developers. the team’s combined goal was to produce a valid, reliable, authentic assessment of ict literacy skills. before the iskills assessment produced figure . first page of a sample score report for an individual. the subsequent pages contain additional performance feedback. figure . sample portion of an institutional score report: compari- son between a user-specified reference group and data from the user’s institution. © educational testing service. all rights reserved. © educational testing service. all rights reserved. � information technology and libraries | september � information technology and libraries | september official scores for test­takers, these specialists—both ets and ict literacy experts—subjected the assess­ ment to a variety of review procedures at many stages of development. these reviews ranged from weekly teleconferences with consortium members during the initial development of assessment tasks (january–july ), to small­scale usability studies in which ets staff observed individual students completing assessment tasks (or mockups of assessment tasks), to field trials that mirrored actual test delivery. the usability studies investigated students’ comprehension of the tasks and testing environment as well as the ease of use of the simulated software in the assessment tasks. the field trials provided opportunities to collect performance data and test the automated scoring algorithms. in some cases, ets staff fine­tuned the scoring algorithms (or developed alternatives) when the scores produced were not psychometrically sound, such as when one element of students’ scores was inconsistent with their overall performance. through these reviews and field trials, the iskills assessment evolved to its current form, targeting and reporting the performance of individuals who complete the seventy­five­minute assessment. in some cases, feedback from experts and field trial participants led to significant changes. for example, the iskills assess­ ment began in as a two­hour assessment (at that time called the ict literacy assessment), that reported scores only to institutions on the aggregated perfor­ mance of their participating students. some students entering higher education found the assessment excessively difficult, which led to the creation of the easier core level assessment. table outlines the participation volumes for the field trials and test administrations. during each field trial, as well as during the institutional administration, feedback was collected from students on their experience with the test via a brief exit survey. table summarizes some results of the exit survey. student reactions to the test were reasonably consistent: most students enjoyed taking the test and found the tasks realistic. in writ­ ten comments, students taking the institutional assess­ ment found the experience rewarding but exhausting, and thought the amount of reading excessive. student feedback directly influenced the design of the core and advanced level assessments, including the shorter test­ table . chronology of field trials and test administrations date administration approximate no. of students approximate no. of participating institutions july–september field trials for institutional assessment , january–april institutional assessment , may field trials for alternative individual assessment structures november field trials for advanced level individual assessment january–may advanced level individual assessment , february field trials for core level individual assessment april–may core level individual assessment , august–december core level: continuous administration , august–december advanced level: continuous administration , note: items in bold represent “live” test administrations in which score reports were issued to institutions, students, or both. article title | author testing information literacy in digital environments | katz taking time and lighter reading load compared with the institutional assessment. as shown in table (bolded rows), test administra­ tions in and early occurred within set time frames. beginning in august , the core and advanced level assessments switched to continuous testing: instead of a specific testing window, institutions create testing sessions to suit the convenience of their resources and students. the tests are still administered in a proctored lab environment, however, to preserve the integrity of the scores. ■ student performance almost , students at sixty­three institutions par­ ticipated during the first administrations of the core and advanced level iskills assessments between january and may . (some institutions administered both the core and advanced level assessments.) test­takers consisted of , high­school students, community college students, and , four­year college and university stu­ dents. institutions selected students to participate based on their assessment goals. some chose to test students enrolled in a particular course, some recruited a random sample, and some issued an open invitation and offered gift certificates or other incentives. because the sample of students is not representative of all united states institu­ tions nor all higher education students, these results do not necessarily generalize to the greater population of college­age students and should therefore be interpreted with caution. even so, the preliminary results reveal interesting trends in the ict literacy skills of participat­ ing students. overall, students performed poorly on both the core and advanced level, achieving only about half of the possible points on the tests. informally, the data suggest that students generally do not consider the needs of an audience when communicating information. for exam­ table . student feedback from the institutional assessment and individual assessments’ field trials statement % agreeing institutional assessment (n= , ) advanced level field trials (n= ) core level field trials (n= ) i enjoyed taking this test. this test was appropriately challenging. i have never taken a test like this one before. to perform well on this test requires thinking skills as well as technical skills. i found the overall testing interface easy to use (even if the tasks themselves might have been difficult). my performance on this test accurately reflects my ability to solve problems using computers and the internet. i didn’t take this test very seriously. the tasks reflect activities i have done at school, work, or home. the software tools were unrealistic. n/a information technology and libraries | september information technology and libraries | september ple, they do not appear to recognize the value of tailor­ ing material to an audience. regarding the ethical use of information, students tend not to check the “fair use” policies of information on the assessment’s simulated web sites. unless the usage policy (for example, copy­ right information) is very obvious, students appeared to assume that they may use information obtained online. on the positive side, test­takers appeared to recognize that .edu and .gov sites are less likely to contain biased material than .com sites. eighty percent of test­takers correctly completed an organizational chart based on e­mailed personnel information. most test­takers cor­ rectly categorized e­mails and files into folders. and when presented with an unclear assignment, percent of test­takers selected the best question to help clarify the assignment. during a task in which students evaluated a set of web sites: ■ only percent judged the objectivity of the sites cor­ rectly; ■ sixty­five percent judged the authority correctly; ■ seventy­two percent judged the timeliness correctly; and ■ overall, only percent of test­takers uniquely identi­ fied the one web site that met all criteria. when selecting a research statement for a class assign­ ment: ■ only percent identified a statement that captured the demands of the assignment; ■ forty­eight percent picked a reasonable but too broad statement; and ■ eight percent picked statements that did not address the assignment. when asked to narrow an overly broad search: ■ only percent selected the correct revision; and ■ thirty­five percent selected a revision that only mar­ ginally narrowed the search results other results suggest that these students’ ict literacy needs further development: ■ in a web search task, only percent entered mul­ tiple search terms to narrow the results; ■ when constructing a presentation slide designed to persuade, percent used only those points directly related to the argument; ■ only a few test­takers accurately adapted existing material for a new audience; and ■ when searching a large database, only percent of test­takers used a strategy that minimized irrelevant results. ■ validity evidence the goal of the iskills assessment is to measure the ict literacy skills of students—higher scores on the assess­ ment should reflect stronger skills. evidence for this validity argument has been gathered since the earliest stages of assessment design, beginning in august . these documentation and research efforts, conducted at ets and at participating institutions, include: ■ the estimated reliability of iskills assessment scores is . (cronbach alpha), which is a measure of test score consistency across various administrations. this level of reliability is comparable to that of many other respected content­based assessments, such as the advanced placement exams. ■ as outlined earlier, the evidence­centered design approach ensures a direct connection between experts’ view of the domain (in this case, ict literacy), evi­ dence of student performance, design of the tasks, and the means for scoring the assessment (katz et al. ). through the continued involvement of the library community in the form of the ict literacy national advisory committee and development committees, the assessment maintains the endorsement of its con­ tent by appropriate subject­matter experts. ■ in november , a panel of experts (librarians and faculty representing high schools, community colleges, and four­year institutions from across the united states) reviewed the task content and scoring for the core level iskills assessment. after investigat­ ing each of the thirty tasks and their scoring in detail, the panelists strongly endorsed twenty­six of the tasks. four tasks received less strong endorsement and were subsequently revised according to the committee’s recommendations. ■ students’ self­assessments of their ict literacy skills align with their scores on the iskills assessment (katz and macklin ). the self­assessment measures were gathered via a survey administered before the assessment. interestingly, although students’ confidence in their ict literacy skills aligned with their iskills scores, iskills scores did not correlate with the frequency with which students reported per­ forming ict literacy activities. this result supports librarians’ claims that mere frequency of use does not translate to good ict literacy skills, and points article title | author testing information literacy in digital environments | katz to the need for ict literacy instruction (oblinger and hawkins ; rockman ). ■ several other validity studies are ongoing, both at ets and at collaborating institutions. these stud­ ies include using the iskills assessment in pre­post evaluations of educational interventions, detailed comparisons of student performance on the assess­ ment and on more real­world ict literacy tasks, and comparisons of iskills assessment scores and scores from writing portfolios. ■ national ict literacy standards and setting cut scores in october , the national forum on information literacy, an advocacy group for information literacy policy (http://www.infolit.org/), announced the formation of the national ict literacy policy council. the policy coun­ cil—composed of representatives from key policy­making, information­literacy advocacy, education, and workforce groups—has the charter to draft ict literacy standards that outline what students should know and be able to do at different points in their academic careers. beginning in , the council will first review existing standards docu­ ments to draft descriptions for different levels of perfor­ mance (for example, minimal ict literacy, proficient ict literacy), creating a framework for the national ict literacy standards. separate performance levels will be defined for the corresponding target population for the core and advanced assessments. these performance­level descrip­ tions will be reviewed by other groups representing key stakeholders, such as business leaders, healthcare educa­ tors, and the library community. the council also will recruit experts in ict literacy and information­literacy instruction to review the iskills assessment and recommend cut scores corresponding to the performance levels for the core and advanced assess­ ments. (a cut score represents the minimum assessment score needed to classify a student at a given performance level.) the standards­based cut scores are intended to help educators determine which students meet the ict literacy standards and which may need additional instruction or remediation. the council will review these recommended cut scores and modify or accept them as appropriately reflecting national ict literacy standards. ■ conclusions ets’s iskills assessment is the first nationally available measure of ict literacy that reflects the richness of that area through simulation­based assessment. owing to the and testing of more than ten thousand students, there is now evidence consistent with anec­ dotal reports of students’ difficulty with ict literacy despite their technical prowess. the results reflect poor ict literacy performance not only by students within one institution, but across the participating sixty­three high schools, community colleges, and four­year colleges and universities. the iskills assessment answers the call of the international ict literacy panel and should inform ict literacy instruction to strengthen these criti­ cal twenty­first­century skills for college students and all members of society. ■ acknowledgments i thank karen bogan, dan eignor, terry egan, and david williamson for their comments on earlier drafts of this article. the work described in this article represents con­ tributions by the entire iskills team at educational testing service and the iskills national advisory committee. works cited american library association. . presidential committee on information literacy: final report. chicago: ala. available online at http://www.ala.org/acrl/legalis.html (accessed june , ). brasley, s. s. . building and using a tool to assess info and tech literacy. computers in libraries , no. : – , – . breivik, p. s. . st century learning and information literacy. change , no. : – . dunn, k. . assessing information literacy skills in the cali­ fornia state university: a progress report. journal of academic librarianship , no. / : – . international ict literacy panel. . digital transformation: a framework for ict literacy. princeton, n.j.: educational testing service. available online at http://www.ets.org/media/ tests/information_and_communication_technology_lit­ eracy/ictreport.pdf (accessed june , ). katz, i. r. . beyond technical competence: literacy in infor­ mation and communication technology. educational technol- ogy magazine , no : – . katz, i. r., and a. macklin. . information and communica­ tion technology (ict) literacy: integration and assessment in higher education. in proceedings of the th international conference on education and information systems, technologies, and applications, f. malpica, a. tremante, and f. welsch, eds. caracas, venezuela: international institute of informatics and systemics. katz, i. r., et al. . assessing information and communications technology literacy for higher education. paper presented at the information technology and libraries | september information technology and libraries | september annual meeting of the international association for educa­ tional assessment, philadelphia, pa. middle states commission on higher education. . develop- ing research and communication skills: guidelines for information literacy in the curriculum. philadelphia: middle states com­ mission on higher education. mislevy, r. j., l. s. steinberg, and r. g. almond. . on the structure of educational assessments. measurement: interdisci- plinary research and perspectives : – . oblinger, d. g., and b. l. hawkins. . the myth about stu­ dent competency. educause review , no. : – . oblinger, d. g., and j. l. oblinger, eds. . educating the net generation. washington, d.c.: educause, http://www. educause.edu/educatingthenetgen (accessed dec. , ). partnership for st century skills. . learning for the st cen- tury: a report and mile guide for st century skills. washington, d.c.: partnership for st century skills. rockman, i. f. . strengthening connections between infor­ mation literacy, general education, and assessment efforts. library trends , no. : – . ———. . introduction: the importance of information lit­ eracy. in integrating information literacy into the higher education curriculum: practical models for transformation. i. f. rockman and associates, eds. san francisco: jossy­bass. the california state university. . information competence initiative web site. http://calstate.edu/ls/infocomp.shtml (accessed june , ). university of central florida. . information fluency initiative web site. http://www.if.ucf.edu/ (accessed june , ). western association of schools and colleges. . handbook of accreditation. alameda, calif.: western association of schools and colleges. available online at http://www.wascsenior .org/wasc/doc_lib/ % handbook.pdf (accessed dec. , ). author id box for column layout this article examines the linguistic structure of folk- sonomy tags collected over a thirty-day period from the daily tag logs of del.icio.us, furl, and technorati. the tags were evaluated against the national information standards organization (niso) guidelines for the con- struction of controlled vocabularies. the results indicate that the tags correspond closely to the niso guidelines pertaining to types of concepts expressed, the predomi- nance of single terms and nouns, and the use of recog- nized spelling. problem areas pertain to the inconsistent use of count nouns and the incidence of ambiguous tags in the form of homographs, abbreviations, and acro- nyms. with the addition of guidelines to the construc- tion of unambiguous tags and links to useful external reference sources, folksonomies could serve as a power- ful, flexible tool for increasing the user-friendliness and interactivity of public library catalogs, and also may be useful for encouraging other activities, such as informal online communities of readers and user-driven readers’ advisory services. o ne of the most daunting challenges of information management in the digital world is the ability to keep, or refind, relevant information; book­ marking is one of the most popular methods for storing relevant web information for reaccess and reuse (bruce, jones, and dumais ). the rising popularity of social bookmark managers, such as del.icio.us, addresses these concerns by allowing users to organize their bookmarks by assigning tags that reflect directly their own vocabu­ lary and needs. the collection of user­assigned tags is referred to commonly as a folksonomy. in recent years, significant developments have occurred in the creation of customizable user features in public library catalogs. these features offer clients the opportunity to customize their own library web pages and to store items of interest to them, such as book lists. client participation in these interfaces, however, is largely reactive; clients can select items from the catalog, but they have little ability to orga­ nize and categorize these items in a way that reflects their own needs and language. digital document repositories, such as library cata­ logs, normally index the subject of their contents via key­ words or subject headings. traditionally, such indexing is performed either by an authority, such as a librarian or a professional indexer, or is derived from the authors of the documents; in contrast, collaborative tagging, or folkson­ omy, allows anyone to freely attach keywords or tags to content. demspey ( ) and ketchell ( ) recommend that clients be allowed to annotate resources of interest and to share these annotations with other clients with similar interests. folksonomies can thus make significant contributions to public library catalogs by enabling cli­ ents to organize personal information spaces; namely, to create and organize their own personal information space in the catalog. clients find items of interest (items in the library catalog, citations from external databases, external web pages, and so on) and store, maintain, and organize them in the catalog using their own tags. in order to more fully understand these applications, it is important to examine how folksonomies are struc­ tured and used, and the extent to which they reflect user needs not found in existing lists of subject headings. the purpose of this proposed research is thus to examine the structure and scope of folksonomies. how are the tags that constitute the folksonomies structured? to what extent does this structure reflect and differ from the norms used in the construction of controlled vocabular­ ies ,such as library of congress subject headings? what are the strengths and weaknesses of folksonomies (for example, reflect user need, ambiguous headings, redun­ dant headings, and so forth)? this article will examine a selection of tags obtained from three folksonomy sites, del.icio.us (referred to henceforth as delicious), furl, and technorati, over a thirty­day period. the structure of these tags will be examined and evaluated against section of the niso guidelines for the construction of controlled vocabularies (niso ), which looks specifically at the choice and form of terms. ■ definitions of folksonomies folksonomies have been described as “user­created meta­ data . . . grassroots community classification of digital assets” (mathes ). wikipedia ( ) describes a folksonomy as “an internet­based information retrieval methodology consisting of collaboratively generated, open­ended labels that categorize content such as web pages, online photographs, and web links.” the concept of collaboration is attributed commonly to folksonomies (bateman, brooks, and mccalla ; cattuto, loreto, and pietronero ; fichter ; golder and huberman the structure and form of folksonomy tags: the road to the public library catalog louise f. spiteri louise f. spiteri (louise.spiteri@dal.ca) is associate professor at the school of information management, dalhousie university, halifax, nova scotia, canada. this research was funded by the oclc/alise library and information science research grant program. the structure and form of folksonomy tags | spiteri � information technology and libraries | september � information technology and libraries | september ; mathes ; quintarelli ; udell ). thomas vander wal, who coined the term folksonomy, argues, however, that: the definition of folksonomy has become completely unglued from anything i recognize. . . . it is not col­ laborative . . . it is the result of personal free tagging of information and objects (anything with a url) for one’s own retrieval. the tagging is done in a social environment (shared and open to others). the act of tagging is done by the person consuming the informa­ tion” (vanderwal.net ). it may be more accurate, therefore, to say that folk­ sonomies are created in an environment where, although people may not actively collaborate in their creation and assignation of tags, they may certainly access and use tags assigned by others. folksonomies thus enable the use of shared tags. folksonomies are used primarily in social bookmark­ ing sites, such as delicious (http://del.icio.us/) and furl (http://www.furl.net/), which allow users to add sites they like to their personal collections of links, to organize and categorize these sites by adding their own terms, or tags, and to share this collection with other people with the same interests. the tags are used to collocate bookmarks within a user’s collection and bookmarks across the entire system, so, for example, the page http://del.icio.us/tag/blogging will show all bookmarks that are tagged with blogging by any member of the delicious site. ■ benefits of folksonomies quintarelli ( ) and fichter ( ) suggest that folk­ sonomies reflect the movement of people away from authoritative, hierarchical taxonomic schemes that reflect an external viewpoint and order that may not necessarily reflect users’ ways of thinking. “in a social distributed environment, sharing one’s own tags makes for innova­ tive ways to map meaning and let relationships naturally emerge” (quintarelli ). vander wal ( ) adds that “the value in this external tagging is derived from people using their own vocabulary and adding explicit mean­ ing, which may come from inferred understanding of the information/object.” an attractive feature of folksonomies is their inclusive­ ness; they reflect the vocabulary of the users, regardless of viewpoint, background, bias, and so forth. folksonomies may thus be perceived to be a democratic system where everyone has the opportunity to contribute and share tags (kroski ). the development of folksonomies may reflect also the difficulty and expense of applying con­ trolled taxonomies to the web: building, maintaining, and enforcing a sound, controlled vocabulary is often simply too expensive in terms of development time and of the steep learning curve needed by the user of the system to learn the classification scheme (fichter ; kroski ; quintarelli ; shirky ). a further limitation of taxonomies is that they may become outdated easily. new concepts or products may emerge that are not yet included in the taxonomy; in comparison, folksonomies easily accommodate such new concepts (fichter ; mitchell ; wu, zubair, and maly, ). shirky ( ) points out that the advantage of folksonomies is not that they are better than controlled vocabularies, but that they are better than nothing. folksonomies follow desire lines, which are expres­ sions of the direct information needs of the user (kroski ; mathes ; merholz ). these desire lines also may reflect the needs of communities of interest: tag­ gers who use same set of tags have formed a group and can seek each other out using simple search techniques. “tagging provides users an easy, yet powerful method to express themselves within a community” (szekely and torres ). ■ weaknesses of folksonomies folksonomies share the problems inherent to all uncon­ trolled vocabularies, such as ambiguity, polysemy, syn­ onymy, and basic level variation (fichter ; golder and huberman ; guy and tomkin ; mathes ). the terms in a folksonomy may have inherent ambiguity as different users apply terms to documents in different ways. the polysemous tag port could refer to a sweet fortified wine, a porthole, a place for loading and unloading ships, the left­hand side of a ship or air­ craft, or a channel endpoint in a communications system. folksonomies do not include guidelines for use or scope notes. folksonomies provide for no synonym control; the terms mac, macintosh, and apple, for example, are all used to describe apple macintosh computers. similarly, both singular and plural forms of terms appear (for example, flower and flowers), thus creating a number of redun­ dant headings. the problem with basic level variation is that related terms that describe an item vary along a continuum of specificity ranging from very general to very specific, so, for example, documents tagged perl and javascript may be too specific for some users, while a document tagged programming may be too general for others. folksonomies provide no formal guidelines for the choice and form of tags, such as the use of com­ pound headings, punctuation, word order, and so forth; for example, should one use the tag vegan cooking or cooking, vegan? guy and tomkin ( ) provide some general suggestions for tag selection best practices, such as the use of plural rather than singular forms, the use article title | author the structure and form of folksonomy tags | spiteri of underscore to join terms in a multiterm concept (for example, open_source), following conventions estab­ lished by others, and adding synonyms. these sugges­ tions are rather too vague to be of much use, however; for example, under what circumstances should singular forms be used (such as noncount nouns), and how should synonyms be linked? ■ applications of folksonomies other than social bookmarking sites, folksonomies are used in commercial shopping sites, such as amazon (http://www.amazon.com/), where clients tag items of interest; these tags can be accessed by people with similar interests. platial (http://www.platial.com/ splash) is used to tag personal collections of maps. examples of the use of folksonomies for intranets include ibm’s social bookmarking application dogear, which allows people to bookmark pages within their intranet (http://domino.watson.ibm.com/cambridge/ research.nsf/ d eb a c f db efc / c ee fbcf fb fc ad ?opendocument), and scuttle (http://sourceforge.net/projects/scuttle/), an open­source bookmarking project that can be hosted on web servers for free. penntags (http://tags.library. upenn.edu/) is a social bookmarking service offered by the university of pennsylvania library to its community members. steve museum is a project that is investigating the incorporation of folksonomies into museum catalogs (trant and wyman ). another potential application of folksonomies is to public library catalogs, where users can organize and tag items of interest in user­specific folders; users could then decide whether or not to post the tags publicly (spiteri ). ■ analyses of folksonomies analysis of the structure, or composition, of tags has thus far been limited; there has been more emphasis placed upon the co­occurrence of tags and their frequency of use. cattuto, loreto, and pietronero ( ) applied a stochas­ tic model of user behavior to investigate the statistical properties of tag co­occurrence; their results suggest that users of collaborative tagging systems share universal behaviors. michlmayr ( ) compared tags assigned to a set of delicious bookmarks to the dmoz (http://www. dmoz.org/) taxonomy, which is designed by a commu­ nity of volunteers. the study concluded that there were few instances of overlap between the two sets of terms. mathes ( ) provides an interesting analysis of the strengths and limitations of the structure of delicious and flickr, but does not provide an explanation of the meth­ odology used to derive his observations; it is not clear, for example, for how long he studied these two sites, how many tags he examined, what elements he was looking for, or what evaluative criteria he applied. golder and huberman ( ) conducted an analysis of the structure of collaborative tagging systems, look­ ing at user activity and kinds and frequencies of tags. specifically, golder and huberman looked at what tags delicious members assigned and how many bookmarks they assigned to each tag. this study identified a number of functions tags perform for bookmarks, including iden­ tifying the: ■ subject of the item; ■ format of the item (for example, blog); ■ ownership of the item; and ■ characteristics of the item (for example, funny). while the golder and huberman study provides an important look at tag use, their study is limited in that they examined only one site for a period of four days; their results are an excellent first step in the analysis of tag use, but the narrow focus of their population and sample size means that their observations are not easily generalized. furthermore, this study focuses more on how bookmarks are associated with tags (for example, how many bookmarks are assigned per tag and by whom) rather than at the structural composition of the tags themselves. guy and tonkin ( ) collected a random sampling of tags from delicious and flickr to see whether “popular objections to folksonomic tagging are based on fact.” the authors do not explain, however, over what period the tags were acquired (for example, over a one­day period, over a month), nor to they provide any evaluative criteria. the tags were entered into aspell, an open source spell checker, from which the authors concluded that percent of flickr and percent of delicious tags were either mis­ spelled, encoded in a manner not understood by aspell, or consisted of compound words of two or more words. tags did not follow convention in such areas as the use of case or singular versus plural forms. while this study certainly focuses upon the structure of the tags, the bases for the authors’ conclusions are problematic. it is not clear that the use of a spell checker is a sufficient measure of quality. does the spell checker allow for cultural variations in spell­ ing (for example, labor or labour)? how well­recognized and comprehensive is the source vocabulary for this spell checker? furthermore, if a tag does not exist in the spell checker, does this necessarily mean that the tag is incor­ rect? tags may include several neologisms, such as podcast- ing, that may not yet exist in conventional dictionaries but are well­recognized in a particular domain. the authors do not mention whether they took into account the cor­ information technology and libraries | september information technology and libraries | september rect use of the singular form of such tags as noncountable nouns (for example, air) or tags that describe disciplines or emotions (for example, history and love). if a named entity (person or organization) was not recognized by aspell, does this mean that the tag was classified as incorrect? lastly, the authors seem to imply that compound words of two or more words are necessarily incorrect, which may not be the case (for example, open source software). the pitfalls of folksonomies have been well­docu­ mented; what is missing is an in­depth analysis of the linguistic structure of tags against an established bench­ mark. while popular opinion suggests that folksonomies suffer from ambiguous and inconsistent structure, the actual extent of these problems is not yet clear; further­ more, analyses conducted so far have not established clear benchmarks of quality pertaining to good tag structure. although there are no guidelines for the construction of tags, recognized guidelines do exist for the construction of terms that are used in taxonomies. although these guidelines discuss the elucidation of inter­term relation­ ships (hierarchical, associative, and equivalent), which does not apply to the flat space of folksonomies, they contain sections pertaining to the choice and formation of concept terms that may, in fact, have relevance for the construction of tags. ■ methodology selection of folksonomy sites tags were chosen from three popular folksonomy sites: delicious, furl, and technorati (http://www.technorati. com/). delicious and furl function as bookmarking sites, while technorati enables people to search for and organize blogs. these sites were chosen because they provide daily logs of the most popular tags that have been assigned by their members on a given day. the daily tag logs from each of the sites were acquired over a thirty­day period (february –march , ). the daily tags for each site were entered into an excel spreadsheet. a list of unique tags for each site was compiled after the thirty­day period; unique refers to the single instance of a tag. some of the tags were used only once during the thirty­day period, while others, such as travel, occurred several times, so travel appears only once in the list of unique tags. variations of the same tag—for example, car or cars, cheney or dick cheney—were considered to constitute two unique tags. only english­language tags were accumulated. the analysis of the tag structure in the three lists was conducted by applying the niso guidelines for thesaurus construction, which are the most current set of recognized guidelines for the: contents, display, construction . . . of controlled vocabu­ laries. this standard focuses on controlled vocabularies that are used for the representation of content objects in knowledge organization systems including lists, syn­ onym rings, taxonomies, and thesauri (niso , ). while folksonomies are not controlled vocabularies, they are lists of terms used to describe content, which means that the niso guidelines could work well as a benchmark against which to examine how folksonomy tags are structured as well as the extent to which this structure reflects the widely accepted norm for controlled vocabu­ laries. section of the guidelines (term choice, scope, and form) was applied to the tags, specifically the following elements (see appendix a for the expanded list): . term choice . grammatical form of terms . nouns . selecting the preferred form only those elements in section that were found to apply to the lists of unique tags are included in appendix a. for each site, the section elements were applied to each unique tag; for example, it was noted whether a tag consists of one or more terms, whether the tag is a noun, adjective, or adverb, and so on. the frequency of occur­ rence of the section elements was noted for each site and then compared across the three sites in order to determine the existence of any patterns in tag structure and the extent to which these patterns reflect current practice in the design of controlled vocabularies. definition and disambiguation of tags the meanings of the tags were determined based upon ( ) the context of their use; and ( ) their definition in three external sources, namely merriam webster online dic­ tionary (http://www.m­w.com/); google (http://www. google.com/); and wikipedia (http://www.wikipedia. org/). merriam­webster was used specifically to define all tags other than those that constitute unique entities (for example, named people, places, organizations, or products) and to determine the various meanings of tags that are homographs (for example, art or web). the actual concept represented by homographs was determined by examin­ ing the sites or blogs to which the tag was assigned. merriam­webster also was used to determine the grammatical form of a tag; for example, noun, verbal noun, adjective, or adverb. determining verbal nouns proved to be complicated, especially given that niso relies only on examples to illustrate such nouns. some tags could serve as both verbal and simple nouns; for example, the tag clipping could describe the activity to clip or an item that has been clipped, such as a newspaper article title | author the structure and form of folksonomy tags | spiteri clipping. similarly, does skiing refer to an activity, or the sport? if the dictionary defined a tag as an activity, the tag was classified as a verbal noun. in the case of tags that were defined as both verbal nouns and simple nouns, the context in which the tag was used determined the final classification. the dictionary also was used to determine the type of concept represented by a tag. the niso guidelines do not define any of these seven types of concepts outlined in section . . ; they provide only a short list of examples for each type. if the term represented by the tag was defined as an activity, property, material, event, discipline or field of study, or unit of measurement, it was classified as such unless the context of the tag suggested otherwise. if none of these six types was defined in the dictionary, the default value of thing was assigned to the tag. these definitions were then compared to the context in which the tag was used. in the case of the tag art, for example, an examination of the sites associated with this tag indicated that it refers to art objects, rather than the discipline, so it was classified as a thing. merriam­webster was used to determine whether a tag constitutes a recognized term in standard english (both united states and united kingdom variants); for example, the tag blogs is a recognized term in the dictionary, while podcasting is not. niso does not provide a clear definition of slang, neologism, or jargon, other than to say that they are nonstandard terms not generally found in dictionaries. is the term podcasting, for example, an instance of slang, jargon, or neologism? at what point does jargon become a neologism? because of the difficulty of distinguishing among these three categories, it was decided to use the broader category nonstandard terms to cover tags that ( ) could not be found in the dictionary; or ( ) are designated as vulgar or slang in the dictionary. google and wikipedia were used to define the mean­ ings of tags that constitute unique entities. wikipedia also was used to distinguish the various meanings of tags that constitute abbreviations or acronyms via its disambigua­ tion pages; for example, the tag nfl is given eight pos­ sible meanings. in this case, the tag nfl is used to refer specifically to the national football league, so the tag is a homograph, noun, and unique entry. ■ tagging conventions and guidelines of the folksonomy sites delicious delicious defines tags as: one­word descriptors that you can assign to your bookmarks. . . . they’re a little bit like keywords but non­hierarchical. you can assign as many tags to a bookmark as you like and easily rename or delete them later. tagging can be a lot easier and more flexible than fitting your information into preconceived categories or folders” (del.icio.us a). the delicious help page for tags encourages people to “enter as many tags as you would like, each separated by a space” in the tag field. this paragraph explains briefly that two lists of tags may appear under the entry form used to enter a bookmark. the first list consists of popular tags assigned by other people to the bookmark in question, while the second consists of recommended tags, which contains a combination of tags that have been assigned by the client in question as well as other users (del.icio.us b). it is not clear how the two lists differ in that they both contain tags assigned by other people to the bookmark at hand. the only tangible guideline provided about how tags should be structured is the sentence “your only limitation on tags is that they must not include spaces.” delicious thus addresses only indirectly the fact that it does not allow multiterm tags; the examples provided suggest ways in which compound terms can be expressed; for example, san­francisco, sanfranciso, san.franciso (del. ico.us b). punctuation thus appears to be allowed in the construction of tags, which is confirmed by the sug­ gestion that asterisks may be used to rate bookmarks: “a tag of * might mean an ok link, *** is pretty good, and a bookmark tagged ***** is awesome” (del.icio.us b). it is thus possible that tags may not consist of recognizable terms, even though asterisks are neither searchable nor indicative of content. furl the furl web site uses the term topics rather than tags, but provides no guidelines or instructions for how to con­ struct these topics. furl mentions only that when entering a bookmark, “a small window will pop up. it should have the title and url of the page you are looking at. enter any additional details (i.e., topic, rating, comments) and click save” (furl ). furl provides all users with a list of default topics to which one can add at will. furl provides no guidelines as to whether single or multiword topics may be used; it is only by trial and error that the user discovers that the latter are, in fact, allowed. technorati in its tags help page, technorati encourages users to “think of a tag as a simple category name. people can categorize their posts, photos, and links with any tag that makes sense” (technorati ). a tag may be “anything, but it should be descriptive. please only use tags that are rel­ evant to the post” (technorati ). technorati tags are � information technology and libraries | september � information technology and libraries | september embedded into individual blogs via the link rel=”tag”; for example: global warming. the tag will appear as simply global warming. no other guidelines are provided about how tags should be constructed. as can be seen, the three folksonomy sites provide very few guidelines or conventions for how tags should be constructed. users are not pointed to the common problems that exist in uncontrolled vocabulary, such as ambiguous headings, homographs, synonyms, spelling variations, and so forth, nor are suggestions made as to the preferred form of tags, such as nouns, plural forms, or the distinction between count nouns (for example, dogs) and mass nouns (for example, air). given this lack of guidance, it is not unreasonable to assume that the tags acquired from these sites will vary considerably in form and structure. ■ findings unless stated otherwise, the number of tags per folk­ sonomy site is for delicious, for furl, and for technorati. homographs the niso guidelines recommend that homographs— terms with identical spellings but different meanings— should be avoided as far as possible in the selection of terms. homographs constitute percent of delicious tags, percent of furl tags, and percent of technorati tags. unique entities constitute a significant proportion of the homographs in all three sites, with percent in delicious, percent in furl, and percent in technorati. the most frequently occurring homographs across the three sites consist predominantly of computer­related terms, such as ajax and css. single-word versus multiword terms the niso guidelines recommend that terms should represent a single concept expressed by a single or mul­ tiword term, as needed. single­term tags constitute percent of delicious tags, percent of furl tags, and percent of technorati tags. the preponderance of single tags in delicious may reflect the fact that it does not allow for the use of spaces between the different elements of the same tag; for example, open source. types of concepts niso provides a list of seven types of concepts that may be represented by terms; while this list is not exhaustive, it represents the most frequently occurring types of con­ cept. table shows the percentage of tags that correspond to each of the seven types of concepts. tags that represent things are clearly predominant in the three sites, with activities and properties forming a distant second and third in importance. none of the tags represent events or measures, and only a fraction of the technorati tags represent materials. the niso guidelines provide no indication of the expected distribution of the types of concepts, so it is difficult to determine to what extent the three folksonomy sites are consistent with other lists of descriptors. none of the tags fell outside the scope of the seven types of concepts. unique entities unique entities may represent the names of people, places, organizations, products, and specific events (niso ). unique entities constitute percent of delicious tags, percent of furl tags, and percent of technorati tags. there is no consistency in the percentage of unique enti­ ties: technorati has nearly twice the percentage of tags than delicious has, and nearly triple the percentage of tags than furl has. computer­related products constitute percent of the unique entities in delicious, percent in furl, and percent in technorati. the remainder of the unique entities in furl and technorati represent places, people, and corporate bodies. the unique entities in technorati are closely related to developments in current news events, an occurrence that is likely due to the site’s focus on blogs rather than web sites. as will be discussed in a subsequent section, the unique entries constitute a significant proportion of the tags that represent ambiguous acronyms or abbreviated terms, such as ajax or psp. table . concepts represented by the tags delicious (%) furl (%) technorati (%) things . materials . activities . events . properties . disciplines . measures . article title | author the structure and form of folksonomy tags | spiteri grammatical forms of terms the niso standards recommend the use of the following grammatical forms of terms: ■ nouns and noun phrases ■ verbal nouns ■ noun phrases ■ premodified noun phrases ■ postmodified noun phrases ■ adjectives ■ adverbs table shows the distribution of the grammatical forms of tags. if all the types of nouns are combined, then percent of delicious tags, percent of furl tags, and percent of technorati tags constitute types of nouns. the gram­ matical structure of the tags in the three folksonomy sites thus reflects very closely the niso recommendations that tags consist of mainly nouns, with the added proviso that adjectives and adverbs be kept to a minimum. none of the folksonomy sites used adverbs as tags, and the num­ ber of adjectives was very small, forming an average total of percent of the tags. nouns (plural and singular forms) niso divides nouns into two categories: count nouns (how many?), and noncount, or mass nouns (how much?). niso recommends that count nouns appear in the plural form and mass nouns in the singular form. niso specifies other types of nouns that appear typi­ cally in the singular form: ■ abstract concepts ■ beliefs; for example, judaism, taoism ■ activities; for example, digestion, distribution ■ emotions; for example, anger, envy, love, pity ■ properties; for example, conductivity, silence ■ disciplines; for example, chemistry, astronomy ■ unique entities table shows the distribution of the singular and plu­ ral forms of noun tags. the term singular nouns was used to collocate all the types of non­plural nouns. table represents the number of tags that constitute count nouns; this does not mean, however, that the tags appeared correctly in the plural form. of the count nouns, percent of delicious tags, percent of furl tags, and percent of technorati tags appeared correctly in the plural form. it should be noted that although table indicates that properties constitute percent of delicious, percent of furl, and percent of technorati tags, most of these tags are adjectives, and thus are not counted in the table. the niso guidelines do not suggest the typical distribution of count versus singular nouns, but table indicates that at least among the three folksonomy sites, singular nouns form the bulk of the tags. table . grammatical form of tags delicious (%) furl (%) technorati (%) nouns verbal nouns noun phrases— premodified noun phrases— postmodified adjectives adverbs table . count and noncount noun tags delicious (%) furl (%) technorati (%) count nouns noncount nouns mass nouns activities properties disciplines unique total information technology and libraries | september information technology and libraries | september spelling the niso guidelines divide the spelling of terms into two sections: warrant and authority. with respect to warrant, niso recommends that “the most widely accepted spell­ ing of words, based on warrant, should be adopted,” with cross­references made between variant spellings of terms. as far as authority is concerned, spelling should follow the practice of well­established dictionaries or glossaries. while spelling refers normally to whole words, i included in this analysis acronyms and abbreviations used to denote unique entities, such as countries or product names, as there are recognized spellings of such acronyms and abbreviations. table shows the tags from the three sites that do not conform to recognized spelling; the terms in italics show the accepted spelling. the number of tags that do not conform to spelling warrant is clearly very few, constituting a total of per­ cent of the delicious tags, percent of the furl tags, and percent of the technorati tags. two of the nonrecognized spellings in delicious are likely due to the difficulty of creating compound tags in this site, as was discussed earlier. the remainder of the tags conformed to recog­ nized spellings as found in the three reference sources consulted. the findings suggest that tags are spelled con­ sistently and in keeping with recognized warrant across the three folksonomy sites. because of the international nature of the three folksonomy sites, no default english spelling was assumed. table shows those tags whose spellings reflect regional variations. none of the three folksonomy sites featured lexical variants of any one tag. as the three sites are united states–based, the preponderance of american spelling is not surprising. what is surprising, however, is that technorati features only the british variants in the total of tags examined in this study. it should be pointed out that the two lexical variants of these terms do appear in the three folksonomy sites; the two variants simply did not appear in the daily logs examined. no system to enable cross­referencing (for example, humour use or see humor) exists in any of the three folksonomy sites, nor is cross­referencing discussed in the help logs of the sites. abbreviations, initialisms, and acronyms niso recommends that the full form of terms should be used. abbreviations or acronyms should be used only when they are so well­established that the full form of the term is rarely used. cross­references should be made between the full and abbreviated forms of the terms. abbreviations and acronyms constitute percent of delicious tags, percent of furl tags, and percent of technorati tags. the majority of these abbreviations and acronyms pertain to unique entities, such as product names (for example, flash, mac, and nfl). in the case of delicious and furl, none of the abbreviated tags is referred to also by its full form. four of the abbreviated technorati tags have full­form equivalents: ■ cheney/dick cheney ■ ie/internet explorer ■ sheehan/cindy sheehan ■ uae/united arab emirates abbreviations and acronyms play a significant role in the ambiguity of the tags from the three sites; they represent percent of the abbreviated delicious tags, percent of the abbreviated furl tags, and percent of the abbreviated technorati tags. furl and technorati are very similar in the proportion of abbreviated tags used, but delicious is significantly higher. the delicious tags are focused more heavily upon computer­related products, which may explain why there are so many more abbrevi­ ated tags, as many of these products are often referred to by these shorter terms; for example, css, flash, apple, and so on. table . tags that do not conform to spelling warrant delicious (n= ) furl (n= ) technorati (n= ) howto (how to) hollywood b- day (hollywood birthday) met-art pics (metropolitan art pictures) opensource (open source) med-books (medical books) superbowl (super bowl) toread (to read) oralsex (oral sex) web- (web . ) table . tags that reflect regional spelling variations delicious (n= ) furl (n= ) technorati (n= ) humor (u.s. spelling) humor (u.s. spelling) favourite (british spelling) jewelry (u.s. spelling) humour (british spelling) article title | author the structure and form of folksonomy tags | spiteri neologisms, slang, and jargon the niso guidelines explain that neologisms, slang, and jargon terms are generally not included in standard dic­ tionaries and should be used only when there is no other widely accepted alternative. nonstandard tags do not constitute a particularly relevant proportion of the total number of tags per site; they account for percent of the delicious tags, percent of the furl tags, and percent of the technorati tags. the nonstandard tags refer almost exclusively to either computer­ or sex­related concepts, such as podcast, wiki, and camsex. nonalphabetic characters this section of the niso guidelines deals with the use of capital letters and nonalphabetic characters. capitalization was not examined in the three folksonomy sites, as none of them are case sensitive; delicious and furl, for exam­ ple, post tags in lower case, regardless of whether the user has assigned upper or lower case, while technorati shows capital letters only if they are assigned by the users themselves. the niso guidelines state that nonalphabetic characters, such as hyphens, apostrophes (unless used for the possessive case), symbols, and punctuation marks, should not be used because they cause filing and search­ ing problems. table shows the occurrence of nonalpha­ betic characters in the three folksonomy sites. a very small proportion of the tags in the three folk­ sonomy sites contains non­alphabetic characters, namely percent of the delicious tags, and percent of the furl and technorati tags. as was discussed previously, the delicious help screens may encourage people to use nonalphabetic characters to construct compound tags; in spite of this, however, such characters are not, in fact, used very frequently. it should be noted that the terms above were all searched, with punctuation intact, in their respective sites; in all three cases, the search engines retrieved the tags and their associated blogs or web sites, which suggests that nonalphabetic characters may not negatively impact searching. ■ discussion and recommendations the tags examined from the three folksonomy sites cor­ respond closely to a number of the niso guidelines pertaining to the structure of terms, namely in the types of concepts expressed by the tags, the predominance of single tags, the predominance of nouns, the use of recognized spelling, and the use of primarily alphabetic characters. potential problem areas in the structure of the tags pertain to the inconsistent use of the singular and plural form of count nouns, the difficulty with creating multi­ term tags in delicious, and the incidence of ambiguous tags in the form of homographs and unqualified abbre­ viations or acronyms. as has been seen, a significant proportion of tags that represent count nouns appears incorrectly in the singular form. because many search engines do not deploy default truncation, the use of the singular or plural form could affect retrieval; a search for the tag computer in delicious, for example, retrieved , hits, while one for computers retrieved , hits. some of the results from the two searches overlapped, but only if both the singular and plural forms of the tags coexist. it would thus be useful for the help features of the folksonomy sites to explain the difference between count and noncount nouns and to discuss the impact of the form of the noun upon retrieval. while all three sites conform to the niso recommendation that single terms be used whenever possible, some concepts cannot be expressed in this fashion, and thus folksonomy sites should accom­ modate the use of multiterm tags. table . nonalphabetic characters delicious (n= ) furl (n= ) technorati (n= ) hyphens — hollywood b-day; url- project consumer- credit; web- . apostrophes — mom’s medical (possessive) valentine’s day (possessive) underscore safari_export blogger_life — full stop — web . (part of product name) web- . (part of product name) forward slash — — /africa + sign — jcr+ — information technology and libraries | september information technology and libraries | september furl and technorati allow for their use, but make no mention of this feature in their help screens, which means that such tags may be constructed inconsistently—for example, by the insertion of punctuation—where a sim­ ple space between the tags will suffice. as has been seen, delicious does not allow directly for the construction of multiterm tags, and in its instructions it actually promotes inconsistency in how various punctuation devices may be used to conflate two or three separate tags, once again at the detriment of retrieval, as is shown below: opensource: , hits open_source: , hits open.source: , hits delicious should consider allowing for the insertion of spaces between the composite words of a compound tag; without this facility, users may be unaware of how to create compound tags. alternatively, delicious should recommend the use of only one punctuation symbol to conflate terms, such as the underscore. furl and technorati should explain clearly that compound tags may be formed by the simple convention of placing a space between the terms. ambiguous headings constitute the most problematic area in the construction of the tags; these headings take the form of homographs and abbreviations or acronyms. in the case of computer­related product names, it may be safe to assume that in the context of an online environ­ ment it is likely that the meaning of these product names is relatively self­evident. in the case of the tag yahoo, for example, none of the sites or blogs associated with this tag pertained to “a member of a race of brutes in swift’s gulliver’s travels who have the form and all the vices of humans, or a boorish, crass, or stupid person” (merriam­ webster ), but referred consistently to the internet service provider and search engine. on the other hand, the tag ajax was used to refer to asynchronous javascript and xml technology as well as to a number of mainly european soccer teams. given the international audience of these folksonomy sites, it may be unwise to assume that the meanings of these homographs are self­evident. library of congress subject headings often uses parenthetical qualifiers to clarify the meaning of terms— for example, python (computer program language)—even though this goes against niso recommendations. it is unlikely, however, that such use of parentheses will be effective in the folksonomy sites. a search for opera (browser), for example, will likely imply an underlying and boolean operator, which detracts from the pur­ pose and value of the parenthetical qualifier; this was confirmed in a furl search, where the terms opera and browser appeared either immediately adjacent to each other or within the same document. the application of the section of the niso guidelines pertaining to abbreviations and acronyms is particularly difficult, as it is important to balance between using abbre­ viated forms of concepts that are so well­known that the full version is hardly used versus creating ambiguous tags. the fact that abbreviated forms appear so prominently in the daily logs of the three folksonomy sites suggests that the full forms of these tags are, in fact, very well­established. at face value, therefore, many of the abbreviated tags are ambiguous because they can refer to different concepts, but it is questionable whether such tags as css, flash, apple, and rss, for example are, in fact, ambiguous to the users of the sites. the use of the full forms for these tags seems cumbersome, as these concepts are hardly ever referred to in their full form. it could possibly be argued, in fact, that in some cases, the full forms may not be familiar; i may know to what concept rss refers, for example, without knowing the specific words represented by the letters r, s, s. the possible ambiguity of abbreviated forms is com­ pounded by the fact that none of the three folkson­ omy sites allows for cross­references between equivalent terms, which is a standard feature of most controlled vocabularies, for example: nfl/national football league use national football league/used for nfl the help screens of the three sites do not address the notion of ambiguity in the construction of tags: they do not draw people’s attention to the inherent ambigu­ ity of abbreviated forms that may represent more than one concept. the sites also fail to address the fact that abbreviated forms (or any tag, for that matter) may be culturally based, so that while the meaning of nfl may be obvious to north american users, this may not be the case for people who live in other geographic areas. it may be useful for the folksonomy sites to add direct links to an online dictionary and to wikipedia, and to encourage people to use these sites to determine whether their cho­ sen tags may have more than one application or meaning; i had not realized, for example, that rss could represent twenty­three different concepts until i used wikipedia and was led to a disambiguation page. access to these external sources may help users decide which full version of the abbreviation to use in the case of ambiguity. the examination of the structure of the tags pointed to some deficiencies in section of the niso guidelines, specifically its occasional lack of sufficient definition or explanation of some of its recommendations. the guidelines list seven types of concepts that are typically represented by controlled vocabulary terms, but rely only upon a few examples to define the meaning and scope of these concepts. the guidelines thus provide no consistent mechanism by which the creators of terms can assess consistently the types of concepts represented. how, for example, is a discipline to be determined? does the term business represent a discipline if it is a subject area that is taught formally in a post­secondary institute, for article title | author the structure and form of folksonomy tags | spiteri example? is it necessary for a discipline to be recognized as such among a majority of educational institutions? in its examples for events, niso lists holidays and revolutions. it is unclear, however, what level of specificity applies to this concept; would christmas, for example, be considered an event or a unique entity/proper noun (which is listed separately from types of concepts)? it is only later in the guidelines, under the examples provided for unique enti­ ties (for example, fourth of july), that one may assume that a named event should be considered a unique entity. verbal nouns also are difficult to determine based only upon the niso examples, and once again no guidelines are provided to determine whether a noun represents an activity or a thing, or possibly both; for example, skiing or clipping. the lack of clear definitions in niso also appeared in the section pertaining to slang, neologisms, and jargon, which are considered to be nonstandard terms that do not generally appear in dictionaries. as was discussed previ­ ously, it is not clear at what point a jargon term or a slang term becomes a neologism. all of the slang tags found in the three sites (for example, babe) appeared in merriam­ webster, which may serve to make this niso section even more ambiguous. ■ conclusion the most notable suggested weaknesses of folksonomies are their potential for ambiguity, polysemy, synonymy, and basic level variation as well as the lack of consistent guidelines for the choice and form of tags. the examina­ tion of the tags of the three folksonomy sites in light of the niso guidelines suggests that ambiguity and polysemy (such as homographs) are indeed problems in the struc­ ture of the folksonomy tags, although the actual propor­ tion of homographs and ambiguous tags each constitutes fewer than one­quarter of the tags in each of the three folksonony sites. in other words, although ambiguity and polysemy are certainly problematic areas, most of the tags in each of the three sites are unambiguous in their meaning and thus conform to niso recommendations. the help sites of the three folksonomy provide few tangible guidelines for ( ) the construction of tags, which affects the construction of multiterm tags; and ( ) the clear distinction between the singular and plural forms of count versus noncount nouns. as has been shown, the use of the singular or plural forms of terms, as well as the use of punctuation to form multiterm tags, affects search results. a large proportion of the tags in all three sites consists of single terms, which mitigates the impact on retrieval, but the inconsistent use of the singular and plural forms of nouns is indeed significant and thus may have marked effect upon retrieval. synonymy and basic level variation were not examined in this study, but are certainly worthy of further exploration. in other areas, the tags conform closely to the niso guidelines for the choice and form of controlled vocabu­ laries. the tags represent mostly nouns, with very few unqualified adjectives or adverbs. the tags represent the types of concepts recommended by niso and conform well to recognized standards of spelling. most of the tags conform to standard usage; there are few instances of nonstandard usage, such as slang or jargon. in short, the structure of the tags in all three sites is well within the standards established and recognized for the construction of controlled vocabularies. should library catalogs decide to incorporate folkson­ omies, they should consider creating clearly written rec­ ommendations for the choice and form of tags that could include the following areas: ■ the difference between count and noncount nouns, as well as an explanation of how the use of the sin­ gular and plural forms affects retrieval. ■ one standard way in which to construct multiterm tags; for example, the insertion of a space between the component terms, or the use of an underscore between the terms. ■ a link to a recognized online dictionary and to wikipedia to enable users to determine the meanings of terms, to disambiguate amongst homographs, and to determine if the full form would be preferable to the abbreviated form. an explanation of the impact of ambiguous tags and homographs upon retrieval would be useful. ■ an acceptable use policy that would cover areas of potential concern, such as the use of potentially offensive tags, overly graphic tags, and so forth. although such terms were not the focus of this study, their presence was certainly evident in some cases, and would need to be considered in an environment that includes clients of all ages. with the use of such expanded guidelines and links to useful external reference sources, folksonomies could serve as a very powerful and flexible tool for increasing the user­friendliness and interactivity of public library catalogs, and also may be useful for encouraging other activities, such as informal online communities of readers and user­driven readers’ advisory services. works cited bateman, s., c. brooks, and g. mccalla. . collabora- tive tagging approaches for ontological metadata in adaptive e-learning systems. http://www.win.tue.nl/sw­el/ / camera­ready/ ­bateman_brooks_mccalla_swel _ final.pdf (accessed jan. , ). � information technology and libraries | september � information technology and libraries | september bruce, h., w. jones, and s. dumais. . keeping and re-finding information on the web: what do people do and what do they need? seattle: information school. http://kftf.ischool.washington .edu/re­finding_information_on_the_web .pdf (accessed jan. , ). cattuto, c., v. loreto, and l. pietronero. . collaborative tag- ging and semiotic dynamics. http://arxiv.org/ps_cache/cs/ pdf/ / .pdf (accessed jan. , ). del.icio.us. a. del.ico.us/about. http://del.icio.us/about/ (accessed jan. , ). del.icio.us. b. del.ico.us/help/tags. http://del.icio.us/help/ tags (accessed jan. , ). dempsey, l. . the recombinant library: portals and people. journal of library administration , no. : – . fichter, d. . intranet applications for tagging and folkson­ omies. online , no. : – . furl. . how to save a page in furl. http://www.furl.net/ howtosave.jsp (accessed jan. , ). golder, s. a., and b. a. huberman. . usage patterns of col­ laborative tagging systems. journal of information science , no. : – . guy, m., and e. tonkin. . tidying up tags? d-lib magazine , no. . http://www.dlib.org/dlib/jan. /guy/ guy.html (accessed jan. , ). ketchell, d. s. . too many channels: making sense out of portals and personalization. information technology and librar- ies , no. : – . kroski, e. . the hive mind: folksonomies and user-based tag- ging. http://infotangle.blogsome.com/ / / /the­hive ­mind­folksonomies­and­user­based­tagging/ (accessed jan. , ). mathes, a. . folksonomies—ccooperative classification and com- munication through shared metadata. http://www.adammathes .com/academic/computer­mediated­communication/ folksonomies.html (accessed jan. , ). merholz, p. . ethnoclassification and vernacular vocabularies. http://www.peterme.com/archives/ .html (accessed jan. , ). merriam­webster. ( ). yahoo. http://www.m­w.com/ (accessed jan. , ). michlmayr, e. . a case study on emergent semantics in communities. http://wit.tuwien.ac.at/people/michlmayr/ publications/michlmayr_casestudy_on_emergentsemantics _final.pdf (accessed jan. , ). mitchell, r. l. . tag teams wrestle with web content. com- puterworld , no. : . niso. . guidelines for the construction, format, and management of monolingual controlled vocabularies. ansi/niso z . ­ . bethesda, md.: national information standards organization. http://www.niso.org/standards/resources/z ­ ­ .pdf (accessed jan. , ). quintarelli, e. . folksonomies: power to the people. http:// www.iskoi.org/doc/folksonomies.htm (accessed jan. , ). shirky, c. . folksonomy. http://www.corante.com/many/ archives/ / / /folksonomy.php (accessed jan. , ). spiteri, l. f. . the use of folksonomies in public library cata­ logues. the serials librarian , no. : – . szekely, b., and e. torres. . ranking bookmarks and bistros: intelligent community and folksonomy development. http:// torrez.us/archives/ / / /tagrank.pdf. (accessed jan. , ). technorati. . technorati help:tags. http://www.technorati. com/help/tags.html (accessed jan. , ). trant, j., and b. wyman. ( ). investigat- ing social tagging and folksonomy in art muse- ums with steve.museum. http://www.archimuse .com/research/www ­tagging­steve.pdf (accessed jan. , ). udell, j. . collaborative knowledge gardening. http://www. infoworld.com/article/ / / / opstrategic_ .html (accessed jan. , ). vander wal, t. . understanding folksonomy: tagging that works. http://s .amazonaws.com/ presentations/ dconstruct/tagging_in_rw.pdf (accessed jan. , ). vanderwal.net. . folksonomy definition and wikipedia. http:// www.vanderwal.net/random/entrysel.php?blog= (accessed jan. , ). wikipedia. . folksonomy. http://en.wikipedia.org/wiki/ folksonomy (accessed jan. , ). wu, h., m. zubair, and k. maly. . harvesting social knowledge from folksonomies. http://delivery.acm.org/ . / / /p ­wu.pdf (accessed jan. , ). article title | author the structure and form of folksonomy tags | spiteri appendix a: list of niso elements . term form . . single word vs. multiword terms . . types of concepts terms for things and their physical parts terms for materials terms for activities or processes terms for events or occurrences terms for properties or states terms for disciplines or subject fields terms for units of measurement . . unique entities . grammatical forms of terms . . nouns and noun phrases . . . verbal nouns . . . noun phrases . . . . premodified noun phrases . . . . postmodified noun phrases . . adjectives . . adverbs . nouns . . count nouns . . mass nouns . . other types of singular nouns . . . abstract concepts . . . unique entities . . spelling . . . spelling—warrant . . . spelling—authorities . . abbreviations, initialisms, and acronyms . . . preference for abbreviation . . . preference for full form . . . . general use . . . . ambiguity . . neologisms, slang, and jargon . . capitalization and nonalphabetic characters information technology and libraries | september author id box for column layout wikis in libraries matthew m. bejune wikis have recently been adopted to support a variety of collaborative activities within libraries. this article and its companion wiki, librarywikis (http://librarywikis. pbwiki.com/), seek to document the phenomenon of wikis in libraries. this subject is considered within the frame- work of computer-supported cooperative work (cscw). the author identified thirty-three library wikis and developed a classification schema with four categories: ( ) collaboration among libraries ( . percent); ( ) collabo- ration among library staff ( . percent); ( ) collabora- tion among library staff and patrons ( . percent); and ( ) collaboration among patrons ( . percent). examples of library wikis are presented within the article, as is a discussion for why wikis are primarily utilized within categories i and ii and not within categories iii and iv. it is clear that wikis have great utility within libraries, and the author urges further application of wikis in libraries. i n recent years, the popularity of wikis has skyrocketed. wikis were invented in the mid­ s to help facilitate the exchange of ideas between computer programmers. the use of wikis has gone far beyond the domain of com­ puter programming, and now it seems as if every google search contains a wikipedia entry. wikis have entered into the public consciousness. so, too, have wikis entered into the domain of professional library practice. the purpose of this research is to document how wikis are used in librar­ ies. in conjunction with this article, the author has created librarywikis (http://librarywikis.pbwiki.com/), a wiki to which readers can submit additional examples of wikis used in libraries. the article will proceed in three sections. the first section is a literature review that defines wikis and introduces computer­supported cooperative work (cscw) as a context for understanding wikis. the second section documents the author’s research and presents a schema for classifying wikis used in libraries. the third section considers the implications of the research results. ■ literature review what’s a wiki? wikipedia ( a) defines a wiki as: a type of web site that allows the visitors to add, remove, edit, and change some content, typically with­ out the need for registration. it also allows for linking among any number of pages. this ease of interaction and operation makes a wiki an effective tool for mass collaborative authoring. wikis have been around since the mid­ s, though it is only recently that they have become ubiquitous. in , ward cunningham launched the first wiki, wikiwikiweb (http://c .com/cgi/wiki), which is still active today, to facilitate the exchange of ideas among computer program­ mers (wikipedia b). the launch of wikiwikiweb was a departure from the existing model of web communica­ tion ,where there was a clear divide between authors and readers. wikiwikiweb elevated the status of readers, if they so chose, to that of content writers and editors. this model proved popular, and the wiki technology used on wikiwikiweb was soon ported to other online communi­ ties, the most famous example being wikipedia. on january , , wikipedia was launched by larry sanger and jimmy wales as a complementary project for the now­defunct nupedia encyclopedia. nupedia was a free, online encyclopedia with articles written by experts and reviewed by editors. wikipedia was designed as a feeder project to solicit new articles for nupedia that were not submitted by experts. the two services coexisted for some time, but in the nupedia servers were shut down. since its launch, wikipedia has undergone rapid growth. at the close of , wikipedia’s first year of operation, there were , articles in eighteen language editions. as of this writing, there are approximately seven million articles in languages, fourteen of which have more than , articles each. as a sign of wikipedia’s growth, when this manuscript was first submitted four months earlier, there were more than five million articles in languages. author’s note: sources in the previous two para­ graphs come from wikipedia. the author acknowledges the concerns within the academy regarding the practice of citing wikipedia within scholarly works; however, it was decided that wikipedia is arguably an authoritative source on wikis and itself. nevertheless, the author notes that there were changes—insubstantial ones—to the cited wikipedia entries between when the manuscript was first submitted and when it was revised four months later. wikis and cscw wikis facilitate collaborative authoring and can be con­ sidered one of the technologies studied under the domain of cscw. in this section, cscw is explained and it is shown how wikis fit within this framework. cscw is an area of computer science research that considers the application of computer technology to sup­ port cooperative, also referred to as collaborative work. the term was first coined in by irene greif ( ) and matthew m. bejune (mbejune@purdue.edu) is an assistant professor of library science at purdue university libraries. he also is a doctoral student at the graduate school of library and information science, university of illinois at urbana-champaign. article title | author wikis in libraries | bejune paul cashman to describe a workshop they were planning on the support of people in work environments with com­ puters. over the years there have been a number of review articles that describe cscw in greater detail, including bannon and schmidt ( ), rodden ( ), schmidt and bannon ( ), sachs ( ), dourish ( ), ackerman ( ), olson and olson ( ), dix, finlay, abowd, and beale ( ), and shneiderman and plaisant ( ). publication in the field of cscw primarily occurs through conferences. the first conference on cscw was held in in austin, texas. since then, the conference has been held biennially in the united states. proceedings are published by the association for computing machinery (acm, http://www.acm.org/). in , the first european conference on computer supported cooperative work (ecscw) was held in amsterdam. ecscw also is held biennially, in odd­numbered years. ecscw proceedings are published by springer (http://www.ecscw.uni­sie­ gen.de/). the primary journal for cscw is computer supported cooperative work: the journal of collaborative computing. publications also appear within publications of the acm and chi, the conference on human factors in computing. cscw and libraries as libraries are, by nature, collaborative work envi­ ronments—library staff working together and with patrons—and as digital libraries and computer technolo­ gies become increasingly prevalent, there is a natural fit between cscw and libraries. the following researchers have applied cscw to libraries. twidale et al. ( ) pub­ lished a report sponsored by the british library research and innovation centre that examined the role of col­ laboration in the information­searching process to inform how information systems design could better address and support collaborative activity. twidale and nichols ( ) offered ethnographic research of physical collaborative environments—in a university library and an office—to aid the design of digital libraries. they wrote two reviews of cscw as applied to libraries—the first was more com­ prehensive (twidale and nichols ) than the second (twidale and nichols ). sánchez ( ) discussed collaborative environments designed and prototyped for digital library environments. classification of collaboration technologies that facilitate collaborative work are typically classified within cscw across two continua: synchronous versus asynchronous, and co­located versus remote. if put together in a two­by­two matrix, there are four possibilities: ( ) synchronous and co­located (same time, same place); ( ) synchronous and remote (same time, different place); ( ) asynchronous and remote (different time, different place); and ( ) asynchronous and co­located (different time, same place). this classification schema was first proposed by johansen et al. ( ). nichols and twidale ( ) mapped work applications within the realm of cscw in figure . wikis are not present in the figure, but their absence is not an indication that they are not cooperative work technologies. rather, wikis were not yet widely in use at the time cscw was considered by nichols and twidale. the author has added wikis to nichols and twidale’s graphical representation in figure . interestingly, wikis are border­crossers fitting within two quadrants: the upper right—asynchronous and co­located; and the lower right—asynchronous and remote. wikis are asynchronous in that they do not require people to be working together at the same time. they are both co­located and remote in that people working collaboratively may not need to be working in the same place. it is also interesting to note that library technologies also can be mapped using johansen’s schema. nichols and twidale ( ) also mapped this, and figure illus­ trates the variety of collaborative work that goes on within libraries. ■ method in order to to discover the widest variety of wikis used in libraries, the author searched for examples of wikis used in libraries within three areas—the lis literature, the library success wiki, and within messages posted on three professional electronic discussion lists. when examples were found, they were logged and classified according to a schema created by the author. results are presented in the next section. the first area searched was within the lis literature. the author utilized the wilson library literature and figure . classification of cscw applications co-located remote synchronous asynchronous meeting rooms distributed meetings muds and moos shared drawing video conferencing collaborative writing team rooms organizational memory workflow web-based applications collaborative writing � information technology and libraries | september � information technology and libraries | september information science database. there were two main types of articles: ones that argued for the use of wikis in libraries, and ones that were case studies of wikis that had been implemented. the second area searched was within library success: a best practices wiki (http://www.libsuccess.org/) (see figure ), created by meredith farkas, distance learning librarian at norwich university. as the name implies, it is a place for people within the library community to share their success stories. posting to the wiki is open to the public, though registration is encouraged. there are many subject areas on the wiki, including management and leadership, readers’ advisory, reference services, infor­ mation literacy, and so on. there also is a section about collaborative tools in libraries (http://www.libsuccess .org/index.php?title=collaborative_tools_in_libraries), in which examples of wikis in libraries are presented. within this section there is a presentation about wikis made by farkas ( ) titled wiki world (http://www. libsuccess.org/indexphp?title=wiki_world), from which examples were culled. the third area that was searched was professional electronic discussion list messages from web lib, dig_ ref, and libref­l. the web lib electronic discussion list (tennant ) is “for the discussion of issues relating to the creation, management, and support of library­ based world wide web servers, services, and applica­ tions.” the list is moderated by roy tennant and the web lib advisory board and was started in . the dig_ref electronic discussion list is a forum for “people and organizations answering the questions of users via the internet” (webjunction n.d.). the list is hosted by the information institute of syracuse, school of information studies, syracuse university, and was created in . the libref­l electronic discussion list is “a moderated discussion of issues related to reference librarianship (balraj ). established in , it’s operated out of kent state university and moderated by a group of list own­ ers. these three electronic discussion lists were selected for two reasons. first, the author is a subscriber to each electronic discussion list, and prior to the research noted there were messages about wikis in libraries. second, based on the descriptions of each electronic discussion list stated above, the selected electronic discussion lists reasonably covered the discussion of wikis in libraries within the professional library electronic discussion lists. one year of messages, november , , through november , , was analyzed for each list. messages about wikis in libraries were identified through key­ word searches against the author’s personal archive of electronic discussion list messages collected over the figure . classification of cscw applications including wikis co-located remote synchronous asynchronous meeting rooms distributed meetings muds and moos shared drawing video conferencing collaborative writing wikis team rooms wikis organizational memory workflow web-based applications collaborative writing figure . classification of collaborative work within libraries co-located remote synchronous asynchronous personal help reference interview issue of book on loan fact-to-face interactions use of opacs database search video conferencing telephone notice boards post-it notes memos documents for study social information filtering e-mail, voicemail distance learning postal services figure �. library success: a best practices wiki (http://www. libsuccess.org/) article title | author wikis in libraries | bejune years. an alternative method would have been to search the web archive of each list, but the author found it easier to search within his mail client, microsoft outlook. messages with the word “wiki” were found in mes­ sages: in web lib, in dig_ref, and in libref­ l. this approach had high recall, as discourse about wikis frequently included the use of the word “wiki,” though low precision, as there were many results that were not about wikis used in libraries. common false hits included messages about the nature study (giles ) that com­ pared wikipedia to encyclopedia britannica, and messages that included the word “wiki” but were simply refer­ ring to wikis, though not examples of wikis used within libraries. from the list of messages, the author read each message and came up with a much shorter list of thirty­nine messages about wikis in libraries: thirty­two in web lib, three in dig_ref, and four in libref­l. ■ results classification of the results after all wiki examples had been collected, it became clear that there was a way to classify the results. in farkas’s ( ) presentation about wikis, she organized wikis in two categories: ( ) how libraries can use wikis with their patrons; and ( ) how libraries can use wikis for knowledge sharing and collaboration. this schema, while it accounts for two types of collaboration, is not granular enough to represent the types of collaboration found within the wiki examples identified. as such, it became clear that another schema was needed. twidale and nichols ( ) identified three types of collaboration within libraries: ( ) collaboration among library staff; ( ) collaboration between a patron and a member of staff; and ( ) collaboration among library users. their classification schema mapped well to the examples of wikis that were identified; however, it too was not granular enough, as it did not distinguish among col­ laboration between library staff intraorganizationally and extraorganizationally, the two most common types of wiki usage found in the research (see appendix). to account for these types of collaboration, which are common not only to wiki use in libraries but to all professional library prac­ tice, the author modified twidale and nichols schema (see figure ). the improved schema also uniformly represents entities across the categories—library staff and member of staff are referred to as “library staff,” and patrons and library users are referred to as “patrons.” examples of wikis used in libraries for each category are provided to better illustrate the proposed classifica­ tion schema. ■ collaboration among libraries the library instruction wiki (http://instructionwiki .org/main_page) is an example of a wiki that is used for collaboration among libraries (figure ). it appears as though the wiki was originally set up to support library instruction within oregon—it is unclear if this was asso­ ciated with a particular type of library, say academic or public—but now the wiki supports library instruction in general. the wiki is self­described as: a collaboratively developed resource for librarians involved with or interested in instruction. all librarians and others interested in library instruction are welcome and encouraged to contribute. the tagline for the wiki is “stop reinventing the wheel”(library instruction wiki ). from this wiki there figure . four types of collaboration within libraries . collaboration among libraries (extra-organizational) . collaboration among library staff (intra-organizational) . collaboration among library staff and patrons . collaboration among patrons figure . wiki world (http://www.libsuccess.org/index.php?title=wiki _world) information technology and libraries | september information technology and libraries | september is a list of library instruction resources that include the fol­ lowing: handouts, tutorials, and other resources to share; teaching techniques, tips, and tricks; class­specific web sites and handouts; glossary and encyclopedia; bibliography and suggested reading; and instruction­related projects, brainstorms, and documents. within the handouts, tutori­ als, and other resources to share section, the author found a wide variety of resources from libraries across the country. similarly, there were a number of suggestions to be found under the teaching techniques, tips, and tricks section. another example of a wiki used for collaboration among libraries is the library success wiki (http://www .libsuccess.org/), one of the sources of examples of wikis used in this research. adding to earlier descriptions of this wiki as presented in this paper, library success seems to be one of the most frequently updated library wikis and perhaps the most comprehensive in its cover­ age of library topics. ■ collaboration among library staff the university of connecticut libraries’ staff wiki (http:// wiki.lib.uconn.edu/) is an example of a wiki used for col­ laboration among library staff (figure ). this wiki is a knowledge base containing more than one thousand infor­ mation technology services (its) documents. its docu­ ments support the information technology needs of the library organization. examples include answers to com­ monly asked questions, user manuals, and instructions for a variety of computer operations. in addition to being a repository of its documents, the wiki also serves as a portal to other wikis within the university of connecticut libraries. there are many other wikis connected to library units; teams; software applications, such as the libraries ils; libraries within the university of connecticut libraries; and other university of connecticut campuses. the health science library knowledge base, stony brook university (http://appdev.hsclib.sunysb.edu/ twiki/bin/view/main/webhome) is another example of a wiki that is used for collaboration among library staff (figure ). the wiki is described as “a space for the dynamic collaboration of the library staff, and a platform of shared resources” (health sciences library ). on the wiki there are the following content areas: news and announcements; hsl departments; projects; trouble­ shooting; staff training resources, working papers and support materials; and community activities, scholarship, conferences, and publications. ■ collaboration among library staff and patrons there are only a few examples of wikis used for collabora­ tion among library staff and patrons to cite as exemplars. one example is the st. joseph county public library (sjpl) subject guides (http://www.libraryforlife.org/ subjectguides/index.php/main_page), seen in figure . this wiki is a collection of resources and services in print and electronic formats to assist library patrons with subject area searching. as the wiki is published by library staff for public consumption, it has more of a professional feel than wikis from the first two categories. pages have images, and the content is structured to look like a standard web page. though the wiki looks like a web page, there still remain a number of edit links that follow each section of text on the wiki. while these tags bear importance for those editing figure . library instruction wiki (http://instructionwiki.org/) figure �. the university of connecticut libraries’ staff wiki (http:// wiki.lib.uconn.edu/) article title | author wikis in libraries | bejune the wiki—library staff only in this case—they undoubtedly puzzle library patrons who think that they have the ability to edit the wiki when, in fact, they do not. another example of collaboration between library staff and patrons that takes a similar approach is the usc aiken gregg­graniteville library web site (http://library. usca.edu/) in figure . as with the sjpl subject guides, this wiki looks more like a web site than a wiki. in fact, the usc aiken wiki conceals its true identity as a wiki even more so than the sjpl subject guides. the only evidence that the web site is a wiki is a link at the bottom of each page that says “powered by pmwiki.” pmwiki (http:// pmwiki.org/) is a content management system that uti­ lizes the wiki technology on the back end to manage a web site while retaining the look and feel of a standard web site. it seems that the benefits of using a wiki in such a way are shared content creation and management. ■ collaboration among patrons as there are only three examples of wikis used for col­ laboration among patrons, all examples will be high­ lighted in this section. the first example is wiki worldcat (http://www.oclc.org/productworks/wcwiki.htm), sponsored by oclc. wiki worldcat launched as a pilot project in september . the service allows users of open worldcat, oclc’s web version of worldcat, to add book reviews to item records. though this wiki does not have many book reviews in it, even for contemporary bestsellers, it gives a taste for how a wiki could be used to facilitate collaboration among patrons. a second example is the biz wiki from ohio university libraries (http://www.library.ohiou.edu/subjects/ bizwiki/index.php/main_page) (see figure ). the biz wiki is a collection of business information resources avail­ able through ohio university. the wiki was created by chad boeninger, reference and instruction librarian, as an alternate form of a subject guide or pathfinder. what separates this wiki from those in the third category, collaboration among library staff and patrons, is that the wiki is editable by patrons as well as librarians. similarly, butler wikiref (http://www .seedwiki.com/wiki/butler_wikiref) is a wiki that has reviews of reference resources created by butler librarians, faculty, staff, and students (see figure ).figure . health sciences library knowledge base (http://appdev .hsclib.sunysb.edu/twiki/bin/view/main/webhome) figure . usc aiken gregg-graniteville library (http://library.usca .edu/) figure . sjcpl subject guides (http://libraryforlife.org/subject guides/index.php/main_page/) information technology and libraries | september information technology and libraries | september full results thirty­three wikis were identified. two wikis were classi­ fied in two categories each. the full results are available in the appendix. table illustrates how wikis were not uniformly distributed across the four categories: category i had . percent, category ii had . percent, category iii had . percent, and category iv had . percent. nearly percent of all examples were found within categories i and ii. as seen in some of the examples in the previous section, wikis were utilized for a variety of purposes. here is a short list of purposes for which wikis were utilized: for sharing information, supporting association work, collecting soft­ ware documentation, supporting conferences, facilitating librarian­to­faculty collaboration, creating digital reposito­ ries, managing web content, creating intranets, providing reference desk support, creating knowledge bases, creating subject guides, and collecting reader reviews. wiki software utilization is summarized in tables and . mediawiki is the most popular software utilized by libraries ( . percent), followed by unknown ( . percent), pbwiki ( . percent), pmwiki ( . percent), seedwiki ( . percent), twiki ( percent), and xwiki ( percent). if the values for unknown are removed from the totals (table ), mediawiki is utilized in almost half ( . percent) of all library wiki applications. ■ discussion with a wealth of examples of wikis in categories i and ii and a dearth of examples of wikis in categories iii and iv, the library community seems to be more comfortable using wikis to collaborate within the community, but less comfortable using wikis to collaborate with library patrons or to enable collaboration among patrons. the research results pose the questions: why are wikis pre­ dominantly used for collaboration within the library community? and why are wikis minimally used for col­ laborating with patrons and helping patrons to collabo­ rate with one another? why are wikis predominantly used for collaboration within the library community? this is perhaps the easier of the two questions to explain. there is a long legacy of cooperation and collaboration intraorganizationally and extraorganizationally within libraries. one explanation for this is the shared bud­ getary climate within libraries. all too often there are insufficient money, staff, and resources to offer desired levels of service. librarians work together to overcome these barriers. prominent examples include coopera­ tive cataloging, interlibrary lending, and the formation of consortia to negotiate pricing. another explanation can be found in the personal characteristics of library professionals. librarianship is a service profession that consequently attracts service­minded individuals who are interested in helping others, whether they are library patrons or fellow colleagues. a third reason is the role of library associations, such as the international federation of library associations and institutions, the american library association, the special libraries association, and the medical library association, as well as many others at the international, national, state, and local lev­ figure . ohio university libraries biz wiki (http://www.library. ohiou.edu/subjects/bizwiki) figure . butler wikiref (http://www.seedwiki.com/wiki/butler_ wikiref) article title | author wikis in libraries | bejune els, and the work that is done through these associations at annual conferences and throughout the year. libraries use wikis to collaborate intraorganizationally and extra­ organizationally because collaboration is what they do most naturally. why are wikis minimally used for collaborating with patrons and helping patrons to collaborate with one another? the reasons for why libraries are only minimally using wikis to collaborate with patrons and for patron collabora­ tion are more difficult to ascertain. however, due to the untapped potential of using wikis, the proposed answers to this question are more important and may lead to future implementations of wikis in libraries. here are four pos­ sible explanations, some more speculative than others. first, perhaps one of the reasons is the result of the way in which libraries are conceived by library patrons and librarians alike. a strong case can be made for libraries as places of collaborative work, and the author takes this posi­ tion. however, historically libraries have been repositories of information, and this remains a pervasive and difficult concept to change—libraries are frequently seen simply as places to get books. in this scenario, the librarian is a gate­ keeper that a patron interacts with to get a book—that is, if the patron interacts with a librarian at all. it also is worthy to note that the relationship is one­way—the patron needs the assistance of librarian, but not the other way around. viewed in these terms, this is not a collaborative situation. for libraries to use wikis for the purpose of collaborating with library patrons, it might demand the reconceptualiza­ tion of libraries by library patrons and librarians. similarly, this extreme conceptualization of libraries does not con­ sider patrons working with one another, even though it is an activity that occurs formally and informally within libraries, not to mention with the emergence of interdisci­ plinary and multidisciplinary work. if wikis are to be used to facilitate collaboration between patrons, the conceptual­ ization of the library by library patrons and librarians must be expanded. second, there may be fears within the library commu­ nity about authority, responsibility, and liability. libraries have long held the responsibility of ensuring the authority of the bibliographic catalog. if patrons are allowed to edit the library wiki, there is potential for negatively affecting the authority of the wiki and even the perceived author­ ity of the library. likewise, there is potential liability in allowing patrons to post to the library wiki. similar con­ table . software totals wiki software no. % mediawiki . unknown . pbwiki . pmwiki . seedwiki . twiki xwiki total: table . software totals without unknowns wiki software no. % mediawiki . pbwiki . pmwiki . seedwiki . twiki . xwiki . total: . table . classification summary category no. % i: collaboration among libraries . ii: collaboration among library staff . iii: collaboration among library staff and patrons . iv: collaboration among patrons . total: . � information technology and libraries | september � information technology and libraries | september cerns have been raised in the past about other collabora­ tive technologies, such as blogs, bulletin boards, mailing lists, and so on, all aspects of the library . movement. if libraries are fully to realize library . as described by casey and savastinuk ( ), miller ( ), and courtney ( ), these issues must be considered. third, perhaps it is due to a matter of fit. it might be the case that wikis are utilized in categories i and ii and not within categories iii and iv because the tools are better suited to support the types of activities within categories i and ii. consider some of the activities listed earlier: sup­ porting association work, collecting software documenta­ tion, supporting conferences, creating digital repositories, creating intranets, and creating knowledge bases. each of these illustrates a wiki that is utilized for the creation of a resource with multiple authors and readers, tasks that are well­suited to wikis. wikipedia is a great example of a wiki with clear, shared tasks for multiple authors and multiple readers and a sense of persistence over time. in contrast, relationships between library staff and patrons do not typically lead to the shared creation of resources. while it is true that the relationship between patron and librarian in the context of a patron’s research assignment can be collab­ orative depending on the circumstances, authorship is not shared but is possessed by the patron. in addition, research assignments in the context of undergraduate coursework are short­lived and seldom go beyond the confines of a particular course. in terms of patrons working together with other patrons, there is the precedent of group work; however, groups often produce projects or papers that share the characteristics of nongroup research assignments listed above. this, of course, does not mean that wikis are not suitable for collaboration within categories iii and iv, but perhaps the opportunities for collaboration are fewer or that they stretch the imagination of the types and ways of doing collaborative work. fourth, perhaps it is a matter of “not yet.” while the research has shown that libraries are not utilizing wikis in categories iii and iv, this may be because it is too soon. it should be noted that wikis are still new technologies. it might be the case that librarians are experimenting in safer contexts so they will gain experience prior to trying more public projects where their expertise will be needed. if this explanation is true, it is expected that more exam­ ples of wikis in libraries will soon emerge. as they do, the author hopes that all examples of wikis in libraries, new and old, will be added to the companion wiki to this article, librarywikis (http://librarywikis.pbwiki.com/). ■ conclusion it appears that wikis are here to stay, and that their utili­ zation within libraries is only just beginning. this article documented the current practice of wikis used in libraries using cscw as a framework for discussion. the author located examples of wikis in three places: within the lis lit­ erature, on the library success wiki, and within messages from three professional electronic discussion lists. thirty­ three examples of wikis were identified and classified using a classification schema created by the author. the schema has four categories: ( ) collaboration among librar­ ies; ( ) collaboration among library staff; ( ) collaboration among library staff and patrons; and ( ) collaboration among patrons. wikis were used for a variety of purposes, including for sharing information, supporting associa­ tion work, collecting software documentation, supporting conferences, facilitating librarian­to­faculty collaboration, creating digital repositories, managing web content, creat­ ing intranets, providing reference desk support, creating knowledge bases, creating subject guides, and collecting reader reviews. by and large, wikis were primarily used to support collaboration among library staff intraorganiza­ tionally and extraorganizationally, with nearly percent ( . percent and . percent respectively) of the examples so identified, and less so in the support of collaboration among library staff and patrons ( . percent) and col­ laboration among patrons ( . percent). a majority of the examples of wikis utilized the mediawiki software ( . percent). it is clear that there are plenty of examples of wikis utilized in libraries, and more to be found each day. it is at this time that the profession is faced with extending the use of this technology, and it is to the future to see how wikis will continue to be used within libraries. works cited ackerman, mark s. . the intellectual challenge of cscw: the gap between social requirements and technical feasibil­ ity. in human-computer interaction in the new millennium, ed. john m. carroll, – . new york: addison­wesley. balraj, leela, et al. libref­l. kent state university librar­ ies. http://www.library.kent.edu/page/ (accessed june , ). archive is available at this link as well. bannon, liam j., and kjeld schmidt. . cscw: four charac­ ters in search for a context. in studies in computer supported cooperative work. ed. john m. bowers and steven d. benford, – . amsterdam: elsevier. casey, michael e., and laura c. savastinuk. . library . . library journal , no. : – . http://www.libraryjournal. com/article/ca .html (accessed june , ). courtney, nancy. . library . and beyond: innovative technolo- gies and tomorrow’s user (in press). westport, conn.: libraries unlimited. dix, alan, et al. . socio­organizational issues and stake­ holder requirements. in human computer interaction, rd ed., – . upper saddle river, n.j.: prentice hall. dourish, paul. . social computing. in where the action is: the foundations of embodied interaction, – . cambridge, mass: mit pr. article title | author wikis in libraries | bejune farkas, meredith. . wiki world. http://www.libsuccess. org/index.php?title=wiki_world (accessed june , ). giles, jim. . internet encyclopaedias go head to head. nature : – . http://www.nature.com/nature/journal/v / n /full/ a.html (accessed june , ). greif, irene, ed. . computer supported cooperative work: a book of readings. san mateo, calif.: morgan kaufmann publishers. health sciences library, state university of new york, stony brook. . health sciences library knowledge base. http://appdev.hsclib.sunysb.edu/twiki/bin/view/main/ webhome (accessed june , ). johansen, robert, et al. . groupware: computer support for business teams. new york: free press. library instruction wiki. . http://instructionwiki.org/ main_page (accessed june , ). miller, paul. . coming together around library . . d- lib magazine , no. . http://www.dlib.org/dlib/april / miller/ miller.html (accessed june , ). nichols, david m., and michael b. twidale. . com­ puter supported cooperative work and libraries. vine : – . http://www.comp.lancs.ac.uk/computing/research/ cseg/projects/ariadne/docs/vine.html (accessed june , ). olson, gary m., and judith s. olson. . groupware and com­ puter­supported cooperative work. in the human-computer interaction handbook: fundamentals, evolving technologies and emerging applications, ed. julie a. jacko and andrew sears, – . mahwah, n.j.: lawrence erlbaum associates, inc.. rodden, tom t. . a survey of cscw systems. interacting with computers , no. : – . sachs, patricia. . transforming work: collaboration, learn­ ing, and design. communications of the acm : – . sánchez, j. alfredo. . hci and cscw in the context of digi­ tal libraries. in chi ‘ extended abstracts on human fac- tors in computing systems. conference on human factors in computing systems. seattle, wash., mar. –apr. . schmidt, kjeld, and liam j. bannon. . taking cscw seri­ ously: supporting articulation work. computer supported cooperative work , no. / : – . shneiderman, ben, and catherine plaisant. . collaboration. in designing the user interface: strategies for effective human- computer interaction, th ed., – . reading, mass.: addison wesley. tennant, roy. . web lib electronic discussion. webjunc­ tion.org. http://lists.webjunction.org/web lib/ (accessed june , ). archive is available at this link as well. twidale, michael b., et al. . collaboration in physical and digital libraries. report no. , british library research and innovation centre. http://www.comp.lancs.ac.uk/ computing/research/cseg/projects/ariadne/bl/report/ (accessed june , ). twidale, michael b., and david m. nichols. a. using studies of collaborative activity in physical environments to inform the design of digital libraries. technical report cseg/ / , computing department, lancaster university, uk. http://www.comp.lancs.ac.uk/computing/research/cseg/ projects/ariadne/docs/cscw .html (accessed june , ). twidale, michael b., and david m. nichols. b. a survey of applications of cscw for digital libraries. technical report cseg/ / , computing department, lancaster university, uk. http://www.comp.lancs.ac.uk/computing/research/cseg/ projects/ariadne/docs/survey.html (accessed june , ). webjunction. n.d. dig_ref electronic discussion list. http:// www.vrd.org/dig_ref/dig_ref.shtml (accessed june , ). wikipedia. a. wiki. http://en.wikipedia.org/wiki/wiki (accessed april , ). wikipedia. b. wikiwikiweb. http://en.wikipedia.org/ wiki/wikiwikiweb (accessed april , ). information technology and libraries | september information technology and libraries | september appendix. wikis in libraries i = collaboration between libraries ii = collaboration between library staff iii = collaboration between library staff and patrons iv = collaboration between patrons category description location wiki software i library success: a best practices wiki—a wiki capturing library success stories. covers a wide variety of topics. also features a presentation about wikis http://www.libsuccess. org/index.php?title=wiki_world http://www.libsuccess.org/ mediawiki i wiki for school library association in alaska http://akasl.pbwiki.com/ pbwiki i wiki to support reserves direct. free, open­source software for managing academic reserves materials developed by emory university. http://www.reservesdirect.org/ wiki/index.php/main_page mediawiki i sunyla new tech wiki—a place for state university of new york (suny) librarians to share how they are using information technologies to interact with patrons http://sunylanewtechwiki.pbwiki. com/ pbwiki i wiki for librarians and faculty members to collaborate across campuses. being used with distance learning instructors and small groups message from robin shapiro. on [dig_ref] electronic discussion list dated / / . unknown i discusses setting up three wikis in last month: “one to sup­ port a pre­conference workshop, another for behind­the­ scenes conferences planning by local organizers, and one for conference attendees to use before they arrived and during the sessions” ( ). fichter, darlene. . using wikis to support online collaboration in libraries. information outlook , no. : ­ . unknown i unofficial wiki to the american library association annual conference http://meredith.wolfwater.com/ wiki/index.php?title=main_page mediawiki i unofficial wiki to the internet librarian conference http://ili .xwiki.com/xwiki/bin/ view/main/webhome xwiki i wiki for the canadian library association (cla) annual conference http://wiki.ucalgary.ca/page/cla mediawiki i wiki for south carolina library association http://www.scla.org/governance/ homepage pmwiki i wiki set up to support national discussion about institutional repositories in new zealand http://wiki.tertiary.govt.nz/ ~institutionalrepositories pmwiki i the oregon library instruction wiki used for sharing infor­ mation about library instruction http://instructionwiki.org/ mediawiki i personal repositories online wiki environment (prowe)— an online repository sponsored by the open university and the university of leicester that uses wikis and blogs to encourage the open exchange of ideas across communities of practice http://www.prowe.ac.uk/ unknown article title | author wikis in libraries | bejune category description location wiki software i lis wiki—space for collecting articles and general informa­ tion about library and information science http://liswiki.org/wiki/main_page mediawiki i making of modern michigan—a wiki to support a state­wide digital library project http://blog.lib.msu.edu/mmmwiki/ index.php/main_page unknown (behind firewall) i wiki used as a web content editing tool in a digital library initiative sponsored by emory university, the university of arizona, virginia tech, and the university of notre dame http://sunylanewtechwiki.pbwiki .com/ pbwiki ii wiki at suny stony brook health sciences library used as knowledge base http://appdev.hsclib.sunysb.edu/ twiki/bin/view/main/webhome; presentation can be found at: http:// ms.cc.sunysb.edu/% edachase/ wikisinaction.htm twiki ii wiki at york university used internally for committee work. exploring how to use wikis as a way to collaborate with users message from mark robertson. on web lib electronic discussion list dated / / . unknown ii wiki for internal staff use at the university of waterloo. they utilize access control to restrict parts of the wiki to groups message from chris gray. on web lib electronic discussion list dated / / . unknown ii wiki at the university of toronto for internal communica­ tions, technical problems, and as a document repository message from stephanie walker. on libref­l electronic discussion list dated / / . unknown ii wiki used for coordination and organization of portable professor program, which appears to be a collaborative infor­ mation literacy program for remote faculty http://tfpp­committee.pbwiki.com/ pbwiki ii the university of connecticut libraries’ staff wiki which is a repository of information technology services documents http://wiki.lib.uconn.edu/wiki/ main_page mediawiki ii wiki used at binghamton university libraries for staff intranet. features pages for committees, documentation, policies, newsletters, presentations, and travel reports screenshots can be found at http://library.lib.binghamton.edu/ presentations/cil /cil% _wikis.pdf mediawiki ii wiki used at the information desk at miami university described in: withers, rob. “something wiki this way comes.” c&rl news , no. ( ): – . unknown ii use of wiki as knowledge base to support reference service http://oregonstate.edu/~reeset/ rdm/ unknown ii university of minnesota libraries staff web site in wiki form https://wiki.lib.umn.edu/ pmwiki ii wiki used to support the mit engineering and science libraries b­team. the wiki may no longer be active, but is still available http://www.seedwiki.com/wiki/b­ team seedwiki iii a wiki that is subject guide at st. joseph county public library in south bend, indiana http://www.libraryforlife.org/ subjectguides/index.php/main_page mediawiki � information technology and libraries | september � information technology and libraries | september category description location wiki software iii wiki used at the aiken library, university of south carolina as a content management system (cms) http://library.usca.edu/main/ homepage pmwiki iii doucette library of teaching resources wiki—a repository of resources for education students http://wiki.ucalgary.ca/page/ doucette mediawiki iv wiki worldcat (wikid) is an oclc pilot project (now defunct) that allowed users to add reviews to open worldcat records http://www.oclc.org/product­ works/wcwiki.htm unknown iii and iv wikiref lists reviews of reference resources—databases, books, web sites, etc. —created by butler librarians, faculty, staff, and students. http://www.seedwiki.com/wiki/ butler_wikiref; reported in matthies, brad, jonathan helmke, and paul slater. using a wiki to enhance library instruction. indiana libraries , no. ( ): – . seedwiki iii and iv wiki used as a subject guide at ohio university http://www.library.ohiou.edu/sub­ jects/bizwiki/index.php/main_page; presentation about the wiki: http://www.infotoday.com/cil / presentations/c ­ _boeninger .pps mediawiki article title | author author id box for column layout thmanager | lacasta, nogueras-iso, lÓpez-pellicer, muro-medrano, and zarazaga-soria author id box for column layout knowledge organization systems denotes formally repre- sented knowledge that is used within the context of digital libraries to improve data sharing and information retrieval. to increase their use, and to reuse them when possible, it is vital to manage them adequately and to provide them in a standard interchange format. simple knowledge orga- nization systems (skos) seem to be the most promising representation for the type of knowledge models used in digital libraries, but there is a lack of tools that are able to properly manage it. this work presents a tool that fills this gap, facilitating their use in different environments and using skos as an interchange format. u nlike the largely unstructured information avail­ able on the web, information in digital libraries (dls) is explicitly organized, described, and man­ aged. in order to facilitate discovery and access, dl sys­ tems summarize the content of their data resources into small descriptions, usually called metadata, which can be either introduced manually or automatically generated (index terms automatically extracted from a collection of documents). most dls use structured metadata in accor­ dance with recognized standards, such as marc (u.s. library of congress ) or dublin core (iso ). in order to provide accurate metadata without ter­ minological dispersion, metadata creators use different forms of controlled vocabularies to fill the content of typi­ cal keyword sections. this increase of homogeneity in the descriptions is intended to improve the results provided by search systems. to facilitate the retrieval process, the same vocabularies used to create the descriptions are usu­ ally used to simplify the construction of user queries. as there are many different schemas for modeling controlled vocabularies, the term knowledge organiza- tion systems (kos) is intended to encompass all types of schemas for organizing information and promoting knowledge management. as hodge ( ) says, “a kos serves as a bridge between the users’ information need and the material in the collection.” some types of kos can be highlighted. examples of simple types are glossaries, which are only a list of terms (usually with definitions), and authority files that control variant ver­ sions of key information (such as geographic or personal names). more complex are subject headings, classifica­ tion schemes, and categorization schemes (also known as taxonomies) that provide a limited hierarchical structure. at a more complex level, kos includes thesauri and less traditional schemes, such as semantic networks and ontologies, that provide richer semantic relations. there is not a single kos on which everyone agrees. as lesk ( ) notes, while a single kos would be advantageous, it is unlikely that such a system will ever be developed. culture constrains the knowledge classifi­ cation scheme because what is meaningful to one area is not necessarily meaningful to another. depending on the situation, the use of one or another kos has its advan­ tages and disadvantages, each one having its place. these schemas, although sharing many characteristics, usually have been treated heterogeneously, leading to a variety of representation formats to store them. thesauri are an example of the format heterogeneity problem. according to iso­ (norm for monolingual thesauri) (iso ), a thesaurus is a set of terms that describe the vocabulary of a controlled indexing language, formally organized so that the a priori relationships between con­ cepts (for example, synonyms, broader terms, narrower terms, and related terms) are made explicit. this stan­ dard is complemented with iso­ (iso ), which describes the model for multilingual thesauri, but none of them describe a representation format. the lack of a stan­ dard representation model has caused a proliferation of incompatible formats created by different organizations. so each organization that wants to use several external thesauri has to create specific tools to transform all of them to the same format. in order to eliminate the heterogeneity of represen­ tation formats, the w c initiative has promoted the development of simple knowledge organization systems (skos) (miles et al. ) for its use in the semantic web environment. skos has been created to represent simple kos, such as subject heading lists, taxonomies, classifica­ tion schemes, thesauri, folksonomies, and other types of controlled vocabulary as well as concept schemes embed­ ded in glossaries and terminologies. although skos has been recently proposed, the number and importance of organizations involved in its creation process (and that publish their kos in this format) indicates that it will probably become a standard for kos representation. skos provides a rich, machine­readable language that is very useful to represent kos, but nobody would expect to have to create it manually or by just using a general­purpose resource description framework (rdf) editor (skos is rdf­based). however, in the digital library area, there are not specialized tools that are able to manage it adequately. therefore, this work tries to fill this gap, describing an open source tool, thmanager, that thmanager: an open source tool for creating and visualizing skos javier lacasta, javier nogueras-iso, francisco javier lópez-pellicer, pedro rafael muro-medrano, and francisco javier zarazaga-soria javier lacasta (jlacasta@unizar.es) is assistant professor, javier nogueras-iso (jnog@unizar.es) is assistant professor, francisco javier lópez-pellicer (fjlopez@unizar.es) is research fellow, pedro rafael muro-medrano (prmuro@ unizar.es) is associate professor, and francisco javier zarazaga-soria (javy@unizar.es) is associate professor in the computer science and systems engineering department, university of zaragoza, spain. � information technology and libraries | september � information technology and libraries | september facilitates the construction of skos­based kos. although thmanager has been created to manage thesauri, it also is appropriate to create and manage any other models that can be represented using skos format. this article describes the thmanager tool, highlight­ ing its characteristics. thmanager’s layer­based architec­ ture permits the reuse of the components created for the management of thesauri in other applications where they are also needed. for example, it facilitates the selection of values from a controlled vocabulary in a metadata cre­ ation tool, or the construction of user queries in a search client. the tool is distributed as open source software accessible through the sourceforge platform (http:// thmanager.sourceforge.net/). ■ state of the art in thesaurus tools and representation models the problem of creating appropriate content for thesauri is of interest in the dl field and other related disciplines, and an increasing number of software packages have appeared in recent years for constructing thesauri. for instance, the web site of willpower information (http://www .willpower.demon.co.uk/thessoft.htm) offers a detailed revision of more than forty tools. some are only avail­ able as a module of a complete information storage and retrieval system, but others also allow the possibility of working independently of any other software. among these thesaurus creation tools, one may note the follow­ ing products: ■ bibliotech (http://www.inmagic.com/). this is a multiplatform tool that forms part of bibliotech pro integrated library system and can be used to build an ansi/niso standard thesaurus (standard z . [ansi ]). ■ lexico (http://www.pmei.com/lexico.html). this is a java­based tool that can be accessed and/or manip­ ulated over the internet. thesauri are saved in a text­based format. it has been used by the u.s. library of congress to manage such vocabularies and thesauri as the thesaurus for graphic materials, the global legal information network thesaurus, the legislative indexing vocabulary, and the symbols of american libraries listing. ■ multites (http://www.multites.com/) is a windows­ based tool that provides support for ansi/niso relationships plus user­defined relationships and comment fields for an unlimited number of thesauri (both monolingual and multilingual). ■ termtree (http://www.termtree.com.au/) is a windows­based tool that uses access, sql server, or oracle for data storage. it can import and export trim thesauri (a format used by the towers records information management system [http://www.towersoft.com/]), as well as a defined termtree tag format. ■ webchoir (http://www.webchoir.com/) is a family of client­server web applications that provides dif­ ferent utilities for thesaurus management in multiple dbms platforms. termchoir is a hierarchical infor­ mation organizing and searching tool that enables one to create and search varieties of hierarchical subject categories, controlled vocabularies, and tax­ onomies based on either predefined standards or a user­defined structure, and is then exported to an xml­based format. linkchoir is another tool that allows indexers to describe information sources using terminology organized in termchoir. and seekchoir is a retrieval system that enables users to browse thesaurus descriptors and their references (broader terms, related terms, synonyms, and so on). ■ synaptica (http://www.synaptica.com/) is a client­ server web application that can be installed locally on a client’s intranet or extranet server. thesaurus data is stored in a sql server or oracle database. the application supports the creation of electronic the­ sauri in compliance with the ansi/niso standard. the application allows the exchange of thesauri in csv (comma­separated values) text format. ■ superthes (batschi et al. ) is a windows­based tool that allows the creation of thesauri. it extends the ansi/niso relationships, allowing many pos­ sible data types to enrich the properties of a concept. it can import and export thesauri in xml and tabular format. ■ tematres (hhttp://r .com.ar/tematres/) is a web application specially oriented to the creation of thesauri, but it also can be used to develop web navigation structures or to manage the documentary languages in use. the thesauri are stored in a mysql database. it provides the created thesauri in zthes (tylor ) or in skos format. finally, it must be mentioned that, given that thesauri can be considered as ontologies specialized in organiz­ ing terminology (gonzalo et al. ), ontology editors have sometimes been used for thesaurus construction. a detailed survey of ontology editors can be found in the denny study ( ). all of these tools (desktop or web­based) present some problems in using them as general thesaurus editors. the main one is the incompatibility in the interchange formats that they support. these tools also present integration problems. some are deeply integrated in bigger sys­ tems and cannot easily be reused in other environments because they need specific software components to work article title | author � thmanager | lacasta, nogueras-iso, lÓpez-pellicer, muro-medrano, and zarazaga-soria � (as dbms to store thesauri). others are independent tools (can be considered as general­purpose thesaurus editors), but their architecture does not facilitate their integration within other information management tools. and most of them are not open source tools, so there is no possibility to modify them to improve their functionality. focusing on the interchange format problem, the iso­ standard (norm for multilingual thesauri) is currently undergoing review by iso tc /sc , and it is expected that the new modifications will include a stan­ dard exchange format for thesauri. it is believed that this format will be based on technologies such as rdf/xml. in fact, some initiatives in this direction have already arisen: ■ the adl thesaurus protocol (janée et al. ) defines an xml­ and http­based protocol for access­ ing thesauri. as a result of query operations, portions of the thesaurus encoded in xml are returned. ■ the language independent metadata browsing of european resources (limber) project has published a thesaurus interchange format in rdf (matthews et al. ). this work introduces an rdf representa­ tion of thesauri, which is proposed as a candidate thesaurus interchange format. ■ the california environmental resources evaluation system (ceres) and the nbii biological resources division are collaborating in a thesaurus partnership project (ceres/nbii ) for the development of an integrated environmental thesaurus and a thesau­ rus networking toolset for metadata development and keyword searching. one of the deliverables of this project is an rdf format to represent thesauri. ■ the semantic web advanced development for europe (swad­europe ) project includes the swad­europe thesaurus activity, which has defined the skos, a set of specifications to represent the knowledge organization systems (kos) on the semantic web (thesauri between them). the british standards bs­ (bsi ) and bs­ (bsi ) (equivalent to the international iso­ and iso­ ) also lack a representation format. the british standards institute idt/ / working group is now developing the bs­ standard that will replace them and whose fifth part will describe the exchange formats and protocols for interoperability of thesauri. the objec­ tive of this working group is to promote the standard to iso, to replace the iso­ and iso­ . here, it is important to remark that given the direct involvement of the idt/ / working group with skos development; probably the two initiatives will not diverge. the new representation format will be, if not exactly skos, at least skos­based. taking into account all these circumstances, skos seems to be the most adequate representation model to store thesauri. given that skos is rdf­based, it can be created using any tool that is able to manage rdf (usually used to edit ontologies); for example, swoop (mindswap group ), protégé (noy et al. ), or triple (wielemaker et al. ). the problem with these tools is that they are too complex for editing and visualizing such a simple model as skos. they are thought to create complex ontologies, so they provide too many options not spe­ cifically adapted to the type of relations in skos. in addition, they do not allow an integrated management of collection of thesauri and other types of controlled vocabularies as needed in dl processes (for example, the creation of metadata of resources, or the construction of queries in a search system). ■ skos model skos is a representation model for simple knowledge organization systems, such as subject heading lists, tax­ onomies, classification schemes, thesauri, folksonomies, other types of controlled vocabulary, and also concept schemes embedded in glossaries and terminologies. this section describes the model, providing characteristics, showing the state of development, and indicating the problems found to represent some types of kos. skos was initially developed within the scope of the semantic web advanced development for europe (swad­europe ). swad­e was created to support w c’s semantic web initiative in europe (part of the ist­ programme). skos is based on a generic rdf schema for thesauri that was initially produced by the desire project (cross et al. ), and further developed in the limber project (matthews et al. ). it has been developed as a draft of an rdf schema for thesauri com­ patible with relevant iso standards, and later adapted to support other types of kos. among the kos already published using this new format are gemet (eea ), agrovoc (fao ), adl feature types (hill and zheng ), and some parts of wordnet lexical data­ base (miller ), all of them available on the skos project web page. skos is a collection of three different rdf schema application profiles: skos­core, to store common prop­ erties and relations; skos­mapping, whose purpose is to describe relations between different kos; and skos­ extension, to indicate specific relations and properties only contained in some type of kos. for the first step of the development of the thmanager tool, only the most stable part of skos has been consid­ ered. figure shows the part of skos­core used. the rest of skos­core is still unstable, so its support has been delayed until it is approved. skos­mapping and skos­extension are still in their first steps of develop­ � information technology and libraries | september � information technology and libraries | september ment and are very unstable, so their management in thmanager also has been delayed until the creation of stable versions. in skos­core, a kos (in our case, usually a the­ saurus) consists of a set of concepts (labelled as skos: concept) that are grouped by a concept scheme (skos: conceptscheme). to distinguish between different mod­ els provided, the skos:conceptscheme contains a uri that identifies it, but to describe the model content to humans, metadata following the dublin core standard also can be added. the relation of the concept scheme with the concepts of the kos is done through the skos: hastopconcept relation. this relation points at the most general concepts of the kos (top concepts), which are used as entry points to the kos structure. in skos, each concept consists of a uri and a set of properties and relations to other concepts. among the properties, skos.preflabel and skos.altlabel provide labels for a concept in different languages. the first one is used to show the label that better identifies a concept (for the­ sauri it must be unique). the second one is an alternative label that contains synonyms or spelling variations of the preferred label (it is used to redirect to the preferred label of the concept). the skos concepts also can contain three other properties called skos.scopenote, skos.definition, and skos.example. they contain annotations about the ways to use a concept, a definition, or examples of use in differ­ ent languages. last, the skos.prefsymbol and skos.altsymbol properties are used to provide a preferred or some alter­ native symbols that graphically represent the concept. for example, a graphical representation is very useful to identify the meaning of a mathematical formula. another example is a chemical formula, where a graphical repre­ sentation of the structure of the substance also provides valuable information to the user. with respect to the relations, each concept indicates by means of the skos:inscheme relation in which concept scheme it is contained. the skos.broader and the skos.nar- rower relations are inverse relations used to model the generalization and specialization characteristics present in many kos (including thesauri). skos.broader relates to more general concepts, and skos.narrower to more spe­ cific ones. the skos.related relation describes associative relationships between concepts (also present in many thesauri), indicating that two concepts are related in some way. with these properties and relations, it is perfectly possible to represent thesauri, taxonomies, and other types of controlled vocabularies. however, there is a problem for the representation of classification schemes that provide multiple coding of terms, as there is no place to store this information. under this category, one may find classification schemes such as iso­ (iso ) (iso standard for coding of languages), which proposes different types of alphanumeric codes (for example, two letters and three letters). for this special case, the skos working group proposes the use of the property skos.notation. although this property is not in the skos vocabulary yet, it is expected to be added in future versions. given the need to work with these types of schemes, this property has been included in the thmanager tool. ■ thmanager architecture this section presents the architecture of thmanager tool. this tool has been created to manage thesauri in skos, but it also is a base infrastructure that facilitates the management of thesauri in dls, simplifying their inte­ gration in tools that need to use thesauri or other types of controlled vocabularies. in addition, to facilitate its use on different computer platforms, thmanager has been developed using the java object­oriented language. the architecture of thmanager tool is shown in figure . the system consists of three layers: first, a repository layer where thesauri are stored and identified by means of associated metadata describing them; second, a per­ sistence layer that provides an api for access to thesauri stored in the repository; and third, a gui layer that offers different graphical components to visualize thesauri, to search by their properties, and to edit them in different ways. the thmanager tool is an application that uses the different components provided by the gui layer to allow the user to manage the thesauri. in addition, the layered architecture allows other applications to use some of the visualization components or the method provided by the persistence layer to provide access to thesauri. the main features that have guided the design of these layers have been the following: a metadata­driven design, efficient management of thesauri, the possibility of interrelating thesauri, and the reusability of thmanager figure . skos model article title | author � thmanager | lacasta, nogueras-iso, lÓpez-pellicer, muro-medrano, and zarazaga-soria � components. the following subsections describe these characteristics in detail. metadata-driven design a fundamental aspect in the repository layer is the use of metadata to describe thesauri. thmanager considers metadata of thesauri as basic information in the thesau­ rus management process, being stored in the metadata repository and managed by the metadata manager. the reason for this metadata­driven design is that thesauri must be described and classified to facilitate the selec­ tion of the one that better fits the user needs, allowing the user to search them not only by their name but also by the application domain or the associated geographi­ cal area between others. the lack of metadata makes the identification of useful thesauri (provided by other organizations) difficult, producing a low reuse of them in other contexts. to describe thesauri in our service, a metadata profile based on dublin core has been created. the reason to use dublin core as basis of this profile has been its extensive use in the metadata community. it provides a simple way to describe a resource using very general metadata ele­ ments, which can be easily matched with complex domain­ specific metadata standards. additionally, dublin core also can be extended to define application profiles for specific types of resources. following the metadata pro­ file hierarchy described in tolosana­calasanz et al. ( ), the thesaurus metadata profile refines the definition and domain of dublin core elements as well as includes two new elements (metadata language and metadata identifier) to appropriately identify the metadata records describing a thesaurus. the profile for thesauri has been described using the iemsr format (heery et al. ) and is distributed with the tool. iemsr is an rdf­based format created by the jisc ie metadata schema registry project to describe metadata application profiles. figure shows the metadata created for gemet thesaurus (the resource), expressed as a hedgehog graph (reinterpreta­ tion of rdf triplets: resources, named properties, and values). the purpose of these metadata is not only to sim­ plify the thesaurus location to a user, but also to facilitate the identification of thesauri useful for a specific task in a machine­to­machine communication. for instance, one may be interested only in thesauri that cover a restricted geographical area or have a specific thematic. efficient thesauri storage thesauri vary enormously in size, ranging from hundreds of concepts and properties to millions. so the time spent on load, navigation, and search processes are a functional restriction for a tool that has to manage them. skos is rdf­based, and because reading rdf to extract the con­ tent is a slow process, the format is not appropriate for inner storage. to provide better access time, thmanager transforms skos into a binary format when a new skos is imported. the persistence layer provides a unified access to the thesaurus repository. this layer is used by the gui layer figure . kos manager architecture viewer generatorviewer generator repository concept repository metadata manager concept manager persistence gui disambiguation tool concept core thesaurus persistence manager skos core skos mapping jena api metadata repository thesaurus metadata applications thmanagerthmanager other tools that use thesauri other tools that use thesauri desktop tools that use thesauri other tools that use thesauri other tools that use thesauri other tools that use thesauri other tools that use thesauri desktop tools that use thesauri desktop tools that use thesauri other tools that use thesauri other tools that use thesauri web services that use thesauri other tools that use thesauri other tools that use thesauri other tools that use thesauri other tools that use thesauri web services that use thesauri web services that use thesauri visualization edition search gui manager figure . metadata of gemet thesaurus european topic centre on catalogue of data sources (etc/cds) general multilingual environmental thesaurus dc:title dcterms:alternative gemet dc:creator [ http://www .ulcc.ac.uk/unesco/concept/mt_mt_ . ] science.environmental sciences and engineering [ http://www .ulcc.ac.uk/unesco/concept/mt_ . ] science.pollution, disasters and security [ http://www .ulcc.ac.uk/unesco/concept/mt_ . ] science.natural resources dc:subject dc:subject dc:subject dc:subject gemet was conceived as a "general" thesaurus, aimed to define a common general language, a core of general terminology for the environment dc:description dc:publisher european environment agency (eea) dc:date - - dc:type [ http://iaaa.cps.unizar.es/dctype/concept/ ] text.reference materials.ontology dc:format [ http://iaaa.cps.unizar.es/mimetype/concept/skos ] skos http://www.eionet.eu.int/gemetdc:identifier dc:language en es fr ... iaaa:metadatalanguage en http://iaaa.cps.unizar.es/ontologies/gemetiaaa:metadataidentifier [ http://www .ulcc.ac.uk/unesco/concept/mt_ . ] science.natural sciences [ http://www.eionet.europa.eu ] european environment information and observation network it can be used whenever there is no commercial profitdc:rights dc:relation us environmental protection agency (epa) dc:contributor dc:source [ http://europa.eu/eurovoc ] eurovoc thesaurus european environment agency (eea) dc:creator ... �� information technology and libraries | september �� information technology and libraries | september to access the thesauri, but it also can be employed by other tools that need to use thesauri outside a desktop environment (for example, a thematic search system accessible through the web that requires browsing a thesaurus to facilitate construction of user queries). this layer performs the transformation of skos to the binary format when a thesaurus is imported. the transformation is provided using the jena library, a popular library to manipulate rdf documents that allows storing them in different kinds of repositories (http://jena.sourceforge. net/). jena provides an open model that can be extended with specialized modules to use other ways of storage, making it possible to easily change the storage format system for another that is more efficient if needed. the data structure used is shown in figure . the model is an optimized representation of the information given by the rdf triplets. the concepts map contains the concepts and their associated relations in the form of key­value pairs: the key is a uri identifying a concept; and the value is a relations object containing the properties of the concept. a relations object is a map that stores the properties of one concept in the form of pairs. the keys used for this map are the names of the typical property types in the skos model (for example, narrower or broader). the only special cases for encoding these property types in the proposed data structure occur when they have a language attribute (for example, prefla- bel, definition, or scopenote). in those cases, we propose the use of a [lang] suffix to distinguish the property type for a particular language. for instance, preflabel_en indicates a preflabel property type in english. additionally, it must be noted that the data type of the property values assigned to each key in the relations map varies upon the semantics given to each property type. the data types fall into the following categories: a string for a preflabel property type; a list of strings for altlabel, definition, scope note, and example property types; a uri for a prefsymbol property type; a list of uris for narrower, broader, related, and altsymbol property types; and a list of notation objects for a notation property type. the data type used for notation values is a complex object because there may be different notation types. a notation object consists of type and value attributes. the type attribute is a uri that identifies a particular notation type and qualifies the associated notation value. additionally, and with the objective of increasing the speed of some operations (for example, navigation or search), some optimizations have been added. first, the uris of the top concepts are stored in the topconcepts list. this list contains redundant information, given that those concepts also are stored in the concepts map, but it makes immediate their location. second, to speed up the search of concepts and the drawing of the alphabetic viewer, the translations map has been added. for each language sup­ ported by the thesaurus, this map contains a translationterm object, or list of pairs , ordered by preflabel. it also contains redundant information that allows the immediate creation of the alphabetic viewer for a language, simplifying the search process; as can be seen later, this does not provides a big over­ head in load time. in addition, if no alphabetic viewer and search are needed, this structure can be removed without affecting the hierarchical viewer. this solution has proven to be useful to manage the kind of thesauri we use (they do not sur­ pass , concepts and about , properties), loading them to memory in an average com­ puter in a reasonable time, and allowing immediate navigation and search (see section ). interrelation of thesauri the vast choice of thesauri that are available nowadays implies an undesired effect of content heterogeneity. although a the­ saurus is usually created for a specific application domain, some of the concepts defined in thesauri from different applica­figure �. persistence model …… relations uri uri relations uri uri relations uri uri valuekey …… relations uri uri relations uri uri relations uri uri valuekey <> concepts uriprefsymbol list altsymbol list notation stringpreflabel_[lang] list altlabel_[lang] list definition_[lang] list scopenote_[lang] list example_[lang] list related list broader list narrower valuekey uriprefsymbol list altsymbol list notation stringpreflabel_[lang] list altlabel_[lang] list definition_[lang] list scopenote_[lang] list example_[lang] list related list broader list narrower valuekey <> relations -type : uri -value : string notation …… list narrower valuekey …… list narrower valuekey <> relations … uri uri uri … uri uri uri <> topconcepts … -concept : uri -label : string translationterm …… listfr listes listen valuekey …… listfr listes listen valuekey <> translations article title | author � thmanager | lacasta, nogueras-iso, lÓpez-pellicer, muro-medrano, and zarazaga-soria � tions domains may be equivalent. in order to facilitate cross­domain classification of resources, users would benefit from the possibility of knowing the connections of a thesaurus in their application domain to thesauri used in other domains. however, it is difficult to manually detect the implicit links between those different thesauri. therefore, in order to automatically facilitate these interthesaurus connections, the persistence layer of thmanager tool provides an interrelation function that relates a thesaurus with respect to an upper­level lexical database (the concept core displayed in figure ). the interrelation mechanism is based on the method presented in nogueras­iso, zarazaga­soria, and muro­ medrano ( ). it is an unsupervised disambiguation method that uses the relations between concepts as disam­ biguation context. it applies a heuristic voting algorithm to select the most adequate sense of the used concept core for each thesaurus concept. at the moment, the concept core is the wordnet lexical database. wordnet is a large english lexical database that groups nouns, verbs, adjectives, and adverbs into sets of cognitive synonyms (synsets), each expressing a distinct concept. those synsets are interlinked by means of conceptual­semantic and lexical relations. the interrelation component has been conceived as an independent module that receives a thesaurus as input in skos and returns the relation respect to concept core using an extended version of the skos mapping model (miles and brickley ). this model, as commented before, is a part of skos that allows describing exact, major, and minor mappings between concepts of two different kos (in this case between a thesaurus and the common core). skos mapping is still in an early stage of development and has been extended in order to provide the needed functionality. the base skos mapping provides the map:exactmatch, map:majormatch, and map:minormatch relations to indicate the degree of relation between two concepts. given that the interrelation algorithm cannot ensure that a mapping is percent exact, only the major and minor match properties are used. the algorithm returns a list of pos­ sible mappings with the lexical database for each concept: the one with the highest probability is assigned as major match, and the rest are assigned as minor matches. to store the interrelation probability, skos mapping has been extended by adding a blank node with the liability of the mapping. also, to be able to know which concepts of which thesauri are equivalents to one of the common core, the inverse relations of map:majormatch and map:minormatch have been created. an example of skos mapping can be seen in figure . there, the concept of gemet thesaurus (alloy) is correctly mapped to the wordnet concept number (alloy, metal) with a probability of . percent, an unrelated minor mapping also is found, but it is given a low probability ( . percent). reusability of thmanager components on top of the api layer, the gui layer has been con­ structed. this layer contains several graphical interfaces to provide different types of viewers, searchers, and edi­ tors for thesauri. this layer is used as base for the con­ struction of the thmanager tool. the tool groups a subset of the provided components, relating them to obtain a final user application that allows the management of the stored thesauri, their visualization (navigation by the concept relations), their edition, and their importation and exportation using skos format. the thmanager tool not only has been created as an independent tool to facilitate thesauri management, but also to allow easy integration in tools that need to use thesauri. it has been done by combining the informa­ tion management with specific graphical interfaces in different black­box components. between the provided components, there is a hierarchical viewer, an alphabetic viewer, a list viewer, a searcher, and an editor, but more components can be constructed if needed. the use of the gui layer as a library of reusable graphical components makes it possible to create different tools that are able to manage thesauri with different user requirements with minimum effort, allowing also the integration of this technology in other applications that need controlled vocabularies to improve their functionality. for example, in a metadata creation tool, it can be used to provide the graphical component to select controlled values from thesauri and automatically insert them in the metadata. it also can be used to provide the list of possible values to use in a web search system, or to provide a thesaurus­ based navigation of a collection of resources in an explor­ atory search system. figure shows the integration process of a thesau­ rus visualization component in an external tool. the provided thesaurus components have been constructed following the java beans philosophy (reusable software components that can be manipulated visually in a builder tool), where a component is a black box with methods to read and change its state that can be reused when needed. here, each thesaurus component is a thesaurusbean that can be directly inserted in a graphical application to use its functionality (visualize or edit thesauri) in a very simple way. the thesaurusbeans are provided by the thesaurusbeanmanager that, given the parameters of the thesaurus to visualize and the type of visualization, returns the most adequate component to use. ■ description of thmanager functionality thmanager tool is a desktop application that is able to manage thesauri stored in skos. as regards to the instal­ � information technology and libraries | september � information technology and libraries | september lation requirements, the application requires mbs of free space on the hard disk. with respect to ram and cpu requirements, they depend greatly on the size and the number of thesauri loaded in the tool. considering the number and size of thesauri used as testbed in section , ram consumption ranges from to mbs, and with a ghz cpu (for example, pentium iv), the load times for the bigger thesauri are acceptable. however, if the size of thesauri is smaller, ram and cpu requirements decrease, being able to operate on a computer with just a ghz cpu (for example, pentium iii) and mbs of ram. given that the management of thmanager is meta­ data oriented, the first window in the application shows a table including the metadata records describing all the thesauri stored in the system (figure ). the selection of a record in this table indicates to the rest of the compo­ nents the selected thesaurus. the creation or deletion of thesauri also is provided here. the only operation that can be performed when no record is selected is to import a new thesaurus stored in skos. to import it, the name of the skos file must be provided. the import tool also contains the option to interrelate the imported thesaurus to the concept core. the metadata of the thesaurus are extracted from inside of the skos if they are available, or they can be provided in an associated xml metadata file. if no metadata record is provided, the application generates a new one with minimum information, using as base the name of the skos file. once the user has selected a thesaurus, it can visualize and modify its metadata or content, export it to skos, or, as commented before, delete it. with respect to the metadata describing a thesaurus, a metadata viewer visualizes the metadata in html and a metadata editor allows the editing of metadata following the thesaurus metadata profile described in the metadata­driven design section (figure shows a screenshot of the metadata edi­ tor). different html views can be provided by adding more css files to the application. the metadata editor is customiz­ able. to add or delete metadata elements to the metadata edi­ tor window, it is only neces­ sary to modify the description of the iemsr profile for thesauri included in the application. the main functionality of the tool is to visualize the thesaurus structure, showing all proper­ ties of concepts and allowing the navigation by relations (see figure ). here, different read­only viewers are provided. there is an alphabetic viewer that shows all the concepts ordered by the preferred label in one language. a hierar­ chical viewer provides navigation by broader and nar­ rower relations. additionally, a hypertext viewer shows all properties of a concept and provides navigation by all its relations (broader, narrower, and related) via hyper­ links. finally, there also is a search system that allows the typical searches needed for thesauri (equals, starts with, contains). currently, search is limited to preferred labels in the selected language, but it could be extended to allow searches by other properties, such as synonyms, defini­ tions, or scope notes. figure . skos mapping extension alloy ... . alloy, metal … . map:majormatch iaaa:probability map:majormatch iaaa:hasmajormatch iaaa:hasmajormatch resource property alloy, metal a mixture containing two or more metallic elements or metallic and nonmetallic elements usually fused together or dissolving into each other when molten; "brass is an alloy of zinc and copper" skos:definition map:minormatch iaaa:hasminormatch admixture, alloy map:minormatch iaaa:hasminormatch http://www.eionet.eu.int/ gemet/concept/ rdf:about a rdf:nodeid a . iaaa:probability rdf:nodeid http://wordnet.princeton.edu/ wordnet_ . / rdf:about skos:preflabel alloy skos:preflabel http://wordnet.princeton.edu/ wordnet_ . / the state of impairing the quality or reducing the value of something skos:preflabel skos:definition rdf:about any of a large number of substances having metallic properties and consisting of two or more elements; with few exceptions, the components are usually metallic elements. (source: mgh) skos:definition figure . gui component integration desktop tool thesaurusbeanmanager type: tree, thesaurus: gemet thesaurusbean article title | author � thmanager | lacasta, nogueras-iso, lÓpez-pellicer, muro-medrano, and zarazaga-soria � all of these viewers are synchronized, so the selec­ tion of a concept in one of them produces the selection of the same concept in the others. the layered architec­ ture described previously allows these viewers to be reused in many situations, including other parts of the thmanager tool. for example, in the thesaurus metadata editor described before, the thesaurus viewer is used to facilitate the selection of values for the subject section of metadata. also, in the thesaurus editor shown later, the thesaurus viewer simplifies the selection of a concept related (by some kind of relation) to the selected, and provides a preview of the hierarchical viewer to help to detect wrong relations. the third available operation is to edit the thesaurus structure. here, to create a thesaurus following the skos model, an edition component is provided (see figure ). the graphical interface shows a list with all the concepts created in the selected thesaurus, allowing the creation of new ones (providing their uris) or deletion of selected ones. once a concept has been selected, its properties and relations to other concepts are shown, allowing the creation of new ones and the deletion of others. to facili­ tate the creation of relations between concepts, a selector of concepts (based in the thesaurus viewer) is provided, allowing the user to add related concepts without manu­ ally typing the uri of the associated concept. also, to see if the created thesaurus is correct, a preview of the hier­ archical viewer can be shown, allowing the user to easily detect problems in the broader and narrower relations. with respect to the interrelation functionality, at the moment the mapping obtained is shown in the thesaurus viewers, but the navigation between equivalent concepts of two thesauri must be be done manually by the user. however, a navigation component still under develop­ ment will allow the user to jump from a concept in a the­ saurus to concepts in others that are mapped to the same concept in the common core. as mentioned before, for efficiency, the format used to store the thesauri in the repository is binary, but the inter­ change format used is skos. so a module for thesauri importation and exportation is provided. this module is able to import from and export to skos. in addition, if the thesaurus has been interrelated with respect to the concept core, it is able to export its mapping to the con­ cept core using the extended version of skos mapping above. ■ results of the work this section shows some experiments performed with the thmanager tool for the storage and management of a selected set of thesauri. in particular, this set of thesauri is relevant in the context of the geographic information community. the increasing relevance of geographic infor­ mation for decision­making and resource management in different areas of government has promoted the cre­ ation of geo­libraries and spatial data infrastructures to facilitate distribution and access of geographic informa­ tion (nogueras­iso, zarazaga­soria, and muro­medrano, ). in this context, complex metadata schemes, such as iso­ , have been proposed for a full­detail descrip­ tion of resources. many of the metadata elements in these schemes are either constrained to a selected vocabulary (iso­ for language encoding, iso­ for country codes, and so on), or the user is told to pick a term from the most suitable thesaurus. the problems with this sec­ ond case are that typically the choice for thesauri is quite open, the thesauri are frequently large, and the exchange format of available thesauri is quite heterogeneous. in such a context, the thmanager tool has proven to be very useful to simplify the management of the used thesauri. at the moment, eighty kos between thesauri and other types of controlled vocabulary have been cre­ ated or transformed to skos and managed through this tool. table shows some of them, indicating their names (name column), the number of concepts (nc column), their total number of properties and relations (np and nr columns), and the number of languages in which concept properties are provided (nl column). to give an idea of the cost of loading these structures, the sizes of skos and binary files (ss and sb columns) are provided in kilobytes (kb). additionally, table compares the performance time of thmanager with respect to other tools that load the figure . thesaurus selector figure �. thesaurus metadata editor �� information technology and libraries | september �� information technology and libraries | september thesauri directly from an rdf file using the jena library (time performance has been obtained using a ghz pentium iv processor). for this purpose, three different load times (in seconds) have been computed. the bt column contains the load time of binary files without the cost of creating the gui for the thesauri viewers. the lt column contains the total load time of binary files (including the time of gui creation and drawing). the jt column contains the time spent by a hypothetical rdf­ based editor tool to invoke jena and load in its memory model the rdf skos files (it does not include gui cre­ ation) containing the thesauri. the difference between the bt and lt column shows the time used to draw the gui once the thesauri have been loaded in memory. the difference between bt and jt columns shows the gain in terms of time of using a binary storage instead of a rdf based one. the thesauri shown in the table are the adl feature types thesaurus (adl ftt), the isoc thesaurus of geography (isoc­g), the iso­ , the unesco thesaurus (unesco ), the ogp surveying and positioning committee code lists (epsg) (ogp ), the multilingual agricultural thesaurus (agrovoc), the european vocabulary thesaurus (eurovoc) (eupo ), the european territorial units (spain and france) (etu), and the general multilingual environmental thesaurus (gemet). they have been selected because they have different sizes and can be used to show how the load time evolves with the thesaurus size. among them, gemet and agrovoc can be high­ lighted. although they are provided as skos, they include nonstandard extensions that we have transformed to standard skos relations and properties. eurovoc and unesco are examples of thesauri provided in formats different than skos that we have completely transformed into skos. the former one was in an xml­based format, and the latter used a plain­text format. another thesaurus transformed to skos is the european territorial units, which contains the administrative political units in spain and france. here, the original source was a collection of heterogeneous documents that contained parts of the needed information and have been processed to generate a skos file. some classification schemes also have been trans­ formed to skos, such as the iso­ and the different epsg codes for coordinate reference systems (includ­ ing datums, ellipsoids, and projections). with respect to controlled vocabularies created (by the authors) in skos using the thmanager tool, there is an extended version of the adl feature types that includes a more detailed clas­ sification of features types and different glossaries used for resource classification. figure depicts the comparison of the different load times shown in table with respect to the size of the rdf skos files. the order of the thesauri in the figure is the same as in the table . it can be seen that the time to con­ struct the model using a binary format is almost half the time spent to create the model using a rdf file. in addi­ tion, once the binary model is loaded, the time to generate the gui is not very dependent on thesaurus size. this is possible thanks to the redundant information added to facilitate the access to top concepts and to speed up load­ ing of the alphabetic viewer. this redundant informa­ tion produces an overhead in the load of the model, but without it the drawing time would be much worse, as it would have to generate it on the fly. however, in spite of the improvements, for the larger thesauri considered, the load time starts to be long, given that it includes the load time of all the structure of the thesaurus in memory and the creation of the objects used to manage it quickly when loaded. but, once it is loaded, future accesses are immediate (quicker than . seconds). these accesses include opening it again, navigating by figure . thesaurus concept selector figure . thesaurus concept editor article title | author � thmanager | lacasta, nogueras-iso, lÓpez-pellicer, muro-medrano, and zarazaga-soria � thesaurus relations, changing the visualization language, and searching concepts by their preferred labels. to minimize the load time, thesauri can be loaded in the background when the application is launched, reducing, in that way, the user perception of the load time. another interesting aspect in figure is the peak of the third element. it corresponds with the iso­ classifica­ tion scheme. it has the special characteristic of not having hierarchy and having many notations. these two character­ istics produce a little increase in the model load time, given that the top concepts list contains all the concepts and the notations are more complex than other relations. but most of the time is used to generate the gui of the tree viewer. the tree viewer gets all the concepts that are top terms, and for each one it asks for their preferred labels in the selected language and sorts them alphabetically to show the first level of the tree. this is fast for a few hundred concepts, but not for the , in the iso­ . however, this problem could be easily solved if the metadata contained a descrip­ tion of the type of kos to visualize. if the tool knew that the kos does not have broader and narrower relations, it could use the structures used to visualize the alphabetic list, which are optimized to show all of the kos concepts rapidly, instead of trying to load it as a tree. the persistence approach used has the advantage of not requiring external persistence systems, such as a dbms, and providing rapid access after loading, but it has the drawback of loading all thesauri in memory (in time and space). so, for much bigger thesauri, the use of some kind of dbms would be necessary. if this change were necessary, minimum modifications would be needed (one class). however, if not all the concepts are loaded, the alphabetic viewer (shows all the concepts) would have to be updated (for example, showing the concepts by pages) or it would become too slow to work with it. ■ conclusions this article has presented a tool for managing the the­ sauri needed in a digital library, for creating metadata, and for running search processes using skos as the interchange format. this work revises the tools that are available to edit thesauri, highlighting the lack of a formalized way to exchange thesauri and the difficulty of integrating those tools in other environments. this work selects skos from the available interchange formats for thesauri as the most promising format to become a standard for skos repre­ sentation, and highlights the lack of tools that are able to manage it properly. the thmanager tool is offered as the solution to these problems. it is an open source tool that can manage the­ sauri stored in skos, allowing their visualization and editing. thanks to the layered architecture, its components can be easily integrated in other applications that need to use thesauri or other controlled vocabularies. additionally, the components can be used to control the possible values used in a web search service to facilitate traditional or exploratory searches based on a controlled vocabulary. the performance of the tool is proved through a series of experiments on the management of a selected set of thesauri. this work analyzes the features of this selected set of thesauri and compares the efficiency of this tool with respect to other tools that load the thesauri directly from a rdf file. in particular, it is shown that the internal representation used by thmanager helps to decrease the time spent for the graphical loading of thesauri, facilitating navigation of the thesaurus contents as well as other typical operations, such as sorting or change of visual­ ization language. additionally, it is worth noting that the tool can be used as a library of components to simplify the integration of the­ sauri in other applications that require the use of controlled vocabularies. thmanager has been integrated within the open source catmdedit tool table . sizes of some thesauri and other types of vocabularies name nc np nr nl lt bt jt ss sb adl ftt . . . isoc­g , , , . . . , , iso­ , , . . . , , unesco , , , . . . , , epsg , , . . . , , agrovoc , , , . . . , , eurovoc , , , . . . , , etu , , , . . . , , gemet , , , . . . , , information technology and libraries | september information technology and libraries | september (zarazaga­soria et al. ), a metadata editor tool for the documentation of geographic information resources (metadata compliant with iso geographic informa­ tion metadata standard). the thesaurusbeans provided in thmanager library have been used to facilitate keyword selection for some metadata elements. the thmanager component library also has contributed to the develop­ ment of catalog search systems guided by controlled vocabularies. for instance, it has been used to build a thematic catalog in the sdiger project (zarazaga­soria ). sdiger is a pilot project on the implementa­ tion of the infrastructure for spatial information in europe (inspire) for the development of a spatial data infrastructure to support access to geographic infor­ mation resources concerned with the european water framework directive. thanks to the thmanager compo­ nents, the thematic catalog allows browsing of resources by means of several multilingual thesauri, including gemet, unesco, agrovoc, and eurovoc. future work will enhance the functionalities provided by thmanager. first, the ergonomics will be improved to show connections between different thesauri. currently, these connections can be computed and annotated, but the gui does not allow the user to navigate them. as the base technology already has been developed, only a graphical interface is needed. second, the tool will be enhanced to support data types different from texts (for example, images, documents, or other multimedia sources) for the encoding of concepts’ property values. third, it has been noted that the thesauri concepts can evolve with time. thus, a mechanism for the managing the different ver­ sions of thesauri will be necessary in the future. finally, improvements in usability also are expected. thanks to the component­based design of thmanager widgets (thesaurusbeans), new viewers or editors can be readily created to meet the needs of specific users. ■ acknowledgments this work has been partially supported by the spanish ministry of education and science through the proj­ ects tin ­ and tic ­ ­c ­ from the national plan for scientific research, development, and technology innovation. the authors would like to express their gratitude to juan josé floristán for his support in the technical development of the tool. references american national standards institute (ansi). . guidelines for the construction, format, and management of monolin­ gual thesauri. ansi/niso z . ­ . revision of z . . batschi, wolf­dieter et al. . superthes: a new software for construction, maintenance, and visualisation of mul­ tilingual thesauri. http://www.t­reks.cnr.it/docs/st_ enviroinfo_ .pdf (accessed sept. , ). british standards institute (bsi). . guide to establishment and development of multilingual thesauri. bs . british standards institute (bsi). . guide to establishment and development of monolingual thesauri. bs . ceres/nbii. . the ceres/nbii thesaurus partnership project. http://ceres.ca.gov/thesaurus/ (accessed june , ). cross, phil, dan brickley, and traugott koch. . rdf the­ saurus specification. technical report , institute for learn- ing and research technology. http://www.ilrt.bris.ac.uk/ discovery/ / /rdf­thes/ (accessed june , ). denny, michael. . ontology building: a survey of edit­ ing tools. xml.com. http://xml.com/pub/a/ / / / ontologies.html (accessed june , ). european environment agency (eea). . general multilingual environmental thesaurus (gemet). version . . european environment information and observation network. http:// www.eionet.europa.eu/gemet/rdf (accessed june , ). european union publication office (eupo). . european vocabulary (eurovoc). publications office. http://europa .eu/eurovoc/ (accessed june , ). food and agriculture organization of the united nations (fao). . agriculture vocabulary (agrovoc). agricul­ tural information management standards. http://www.fao. org/aims/ag% alpha.htm (accessed june , ). gonzalo, julio, et al. . applying eurowordnet to cross­lan­ guage text retrieval. computers and the humanities , no. / (special issue on euroword­net): – . heery, rachel, et al. . jisc metadata schema registry. in th acm/ieee-cs joint conference on digital libraries, – . new york: acm pr. hill, linda, and qi zheng. . indirect geospatial referencing through place names in the digital library: alexandria digi­ figure . thesaurus load times skos file size (kb) lo ad t im e (s ) rdf (jena) binary thmanager article title | author thmanager | lacasta, nogueras-iso, lÓpez-pellicer, muro-medrano, and zarazaga-soria tal library experience with developing and implementing gazetteers. in asis ‘ : proceedings of the nd asis annual meeting: knowledge: creation, organization, and use, – . med­ ford, n.j.: information today, for the ameircan society for information science. hodge, gail. . systems of knowledge organization for digital libraries: beyond traditional authority files. washington, d.c.: the digital library federation. international organization for standardization (iso). . guidelines for the establishment and development of multilingual thesauri. iso . international organization for standardization (iso). . guidelines for the establishment and development of monolingual thesauri. iso . international organization for standardization (iso). . codes for the representation of names of languages. iso . international organization for standardization (iso). . information and documentation—the dublin core metadata ele- ment set. iso : . janée, greg, satoshi ikeda, and linda l. hill. . the adl the­ saurus protocol. http://www.alexandria.ucsb.edu/~gjanee/ thesaurus/ (accessed june , ). lesk, michael. . practical digital libraries. san francisco: books, bytes, and bucks. matthews, brian m., et al. . internationalising data access through limber. in third international workshop on interna- tionalisation of products and systems: – . milton keynes (uk). http://epubs.cclrc.ac.uk/bitstream/ /limber_iwips.pdf (accessed june , ). miles, alistair, and dan brickley, eds. . skos mapping vocab­ ulary specification. w c. http://www.w .org/ / / skos/mapping/spec/ ­ ­ .html (accessed june , ). miles, alistair, brian matthews, and michael wilson. . skos core: simple knowledge organization for the web. in dublin core annual conference—vocabularies in practice, – . madrid: universidad carlos ii de madrid. miller, george a. . wordnet: an on­line lexical database. int. j. lexicography : – . mindswap group. . swoop a hypermedia­based feath­ erweight owl ontology editor. maryland information and network dynamics lab. semantic web agents project. http://www.mindswap.org/ /swoop/ (accessed june , ). nogueras­iso, javier, francisco javier zarazaga­soria, and pedro rafael muro­medrano. . geographic information metadata for spatial data infrastructures—resources, interoperability, and information retrieval. new york: springer verlag. noy, natalie f., ray w. fergerson, and mark a. musen. . the knowledge model of protégé : combining interoper­ ability and flexibility. in knowledge engineering and knowledge management: methods, models, and tools: th international con- ference, ekaw , juan-les-pins, france, october – , : proceedings, ­ (lecture notes in computer science, ). new york: springer. ogp surveying & positioning committee. . surveying and positioning. http://www.epsg.org/ (accessed june , ). semantic web advanced development for europe (swad­ europe). . semantic web advanced development for europe thesaurus activity. http://www.w .org/ /sw/ europe/ reports/thes (accessed june , ). tolosana­calasanz, r., et al. . semantic interoperability based on dublin core hierarchical one­to­one mappings. international journal of metadata, semantics, and ontologies , no. : – . tylor, mike. . the zthes specifications for thesaurus rep­ resentation, access, and navigation. http://zthes.z .org/ (accessed june , ). united nations educational, scientific, and cultural organiza­ tion (unesco). . unesco thesaurus: a structured list of descriptors for indexing and retrieving literature in the fields of education, science, social and human science, culture, com- munication and information. paris: unesco publ. u.s. library of congress. network devlopment and marc standards office. . marc standards. http://www.loc. gov/marc/ (accessed june , ). wielemaker, jan, guss schreiber, and bob wielinga . . using triples for implementation: the triple ontology-manipulation tool (lecture notes in computer science, ): – . new york: springer. zarazaga­soria, francisco javier, et al. . a java tool for creating iso/fgdc geographic metadata. in geodaten- und geodienste- infrastukuren—von der forschung zur praktischen anwendung: beitrage ze den münsteraner gi-tagen, / . juni (ifgiprints, ). münster, germany: institut fur geoin­ formatik, universitat münster. zarazaga­soria, francisco javier, et al. . providing sdi ser­ vices in a cross­border scenario: the sdiger project use case. in research and theory in advancing spatial data infra- structure concepts, – . redlands, calif.: esri. ebsco cover lita cover , cover index to advertisers a s i approach the end of my tenure as ital edi­ tor, i reflect on the many lita members who have not submitted articles for possible publica­ tion in our journal. i am especially mindful of the smaller number who have promised or hinted or implied that they intended to or might submit articles. admittedly, some of them may have done so because i asked them, and their replies to me were the polite ones that one expects of the honorable members of the library and information technology association of the american library association. librarians are as individuals almost all or almost always polite in their professional discourse. pondering these potential authors, particularly the smaller number, i conjured a mental picture of a fictional, male, potential ital author. i don’t know why my fic­ tional potential author was male—it may be because more males than females are members of that group; it may be because i’m a male; or it may be unconscious sex­ ism. i’m not very self­analytic. my mental picture of this fictional male potential author saw him driving home from his place of employ­ ment after having an after­work half gallon of rum when, into the picture, a rattlesnake crawled on to the seat of his car and bit him on the scrotum. lucky him: he was, after all, a figment of my imagina­ tion. (any resemblance between my fictional author and a real potential author is purely coincidental.) lucky me: we all know that such an incident is not unthinkable in library land. lucky lita: it is unlikely that any member will cancel his or her membership or any subscriber, his, her, or its subscription because the technical term “scro­ tum” found its way into my editorial. ital is, after all, a technology journal, and members and readers ought to be offended if our journal abjures technical terminology. likewise they should be offended if our articles discuss library technology issues misusing technical terms or concepts, or confusing technical issues with policy issues, or stating technology problems or issues in the title or abstract or introduction then omitting any mention of said problems until the final paragraph(s). ital referees are quite diligent in questioning authors when they think terminology has been used loosely. their close readings of manuscripts have caught more than one author mislabeling policies related to the uses of informa­ tion technologies as if the policies were themselves tech­ nical conundrums. most commonly, they have required authors who state major theses or technology problems at the beginnings of their manuscripts, then all but ignore these until the final paragraphs, to rewrite sections of their manuscripts to emphasize the often interesting questions raised at the outset. what, pray tell, is the editor trying to communicate to readers? two things, primarily. first, i have been following with interest the several heated discussions that have taken place on lita­l for the past number of months. sometimes, the idea of the traditional quarterly scholarly/professional journal in a field changing so rapidly may seem almost quaint. a typical ital article is five months old when it is pub­ lished. a typical discussion thread on lita­l happens in “real time” and lasts two days at most. a small number of participants raise and “solve” an issue in less than a half dozen posts. a few times, however, a question asked or a comment posted by a lita member has led to a flurry of irrelevant postings, or, possibly worse, sustained bomb­ ing runs from at least two opposing camps that have left some members begging to be removed from the list until the all clear signal has been sounded. i’ve read all of these, and i could not help but won­ der, what if ital accepted manuscripts as short as lita­l postings? what would our referees do? i suspect, for our readers’ sakes, most would be rejected. authors whose manuscripts are rejected receive the comments made by the referees and me explaining why we cannot accept their submissions. the most frequent reason is that they are out of scope, irrelevant to the purposes of lita. when someone posts a technology question to lita­l that gener­ ates responses advising the questioner that implementing the technology in question is bad policy, the responses are, from an editor’s point of view, out of scope. how many lita members have authority—real authority—to set policy for their libraries? a second “popular” reason for rejections is that the manuscripts pose “false” problems that may be technological but that are not technologies that are within the “control” of libraries. these are out of scope in a different manner. third, some manuscripts do not pass the “so what” test. some days i wish that lita­l responders would referee, honestly, their own responses for their relevance to the questions or issues or so­whatness and to the membership. second, and more importantly to me, lita members, whether or not your bodies include the part that we all have come to know and defend, do you have the “­” to send your ital editor a manuscript to be chewed upon not by rattlesnakes but by the skilled professionals who are your ital editorial board members and referees? i hope (and do i dare beg again?) so. your journal will not suffer quaintness unless you make it so. editorial: the virtues of deliberation john webb john webb (jwebb@wsu.edu) is a librarian emeritus, washington state university, and editor of information technology and libraries. editorial | webb drawing upon findings from a national survey of u.s. public libraries, this paper examines trends in internet and public computing access in public libraries across states from to . based on library-supplied information about levels and types of internet and public computing access, the authors offer insights into the net- work-based content and services that public libraries pro- vide. examining data from to reveals trends and accomplishments in certain states and geographic regions. this paper details and discusses the data, identi- fies and analyzes issues related to internet access, and suggests areas for future research. t his article presents findings from the and public libraries and the internet studies detail­ ing the different levels of internet access available in public libraries in different states. at this point, . percent of public library branches are connected to the internet and . percent of connected public library branches offer public internet access. however, the types of access and the quality of access available are not uniformly distributed among libraries or among the libraries in various states. while the data at the national level paint a portrait of the internet and public computing access provided by public libraries overall, studies of these differences among the states can help reveal successes and lessons that may help libraries in other states to increase their levels of access. the need to continue to increase the levels and quality of internet and public computing access in public libraries is not an abstract problem. the services and con­ tent available on the internet continue to require greater bandwidth and computing capacity, so public libraries must address ever­increasing technological demands on the internet and computing access that they provide. public libraries are also facing increased external pressure on their internet and computing access. as patrons have come to rely on the availability of internet and computing access in public libraries, so too have government agencies. many federal, state, and local government agencies now rely on public libraries to facilitate citizens’ access to e­government services, such as applying for the federal prescription drug plans, filing taxes, and many other interactions with the gov­ ernment. further, public libraries also face increased demands to supply public access computing in times of natural disasters, such as the major hurricanes of and . as a result, both patrons and govern­ ment agencies depend on the internet and computing access provided by public libraries, and each group has different, but interrelated, expectations of what kinds of access public libraries should provide. however, the data indicate that public libraries are at capacity in meet­ ing some of these expectations, while some libraries lack the funding, technology­support capacity, space, and infrastructure (e.g., power, cabling) to reach the expecta­ tions of each respective group. as public libraries (and the internet and public com­ puting access they provide) continue to fill more social roles and expectations, a range of new ideas and strate­ gies can be considered by public libraries to identify suc­ cessful methods for providing access that is high quality and sufficient to meet the needs of patrons and commu­ nity. the goals of the public libraries and the internet stud­ ies have been to help provide an understanding of the issues and needs of libraries associated with providing internet­based services and resources. the public libraries and the internet study employed a web­based survey approach to gather both quantitative and qualitative data from a sample of the , public library outlets in the united states. a sample was drawn to accurately represent metropolitan status (roughly equating to their designation of urban, suburban, or rural libraries), poverty levels (as derived through census data), state libraries, and the national picture, producing a sample of , public library out­ lets. the survey received a total of , responses for a response rate of percent. the data in this article, unless otherwise noted, are drawn from the and public libraries and the internet studies. while the survey received responses from librar­ ies in all fifty states, there were not enough responses in all states from which to present state­level findings. the study was able to provide state­level analysis for thirty­five states (including washington, d.c.) in and forty­four states at the outlet level (including washington, d.c.) and forty­two states at the system level (including washington, d.c.) in . in addi­ tion, there was some variance in states with adequate responses between the and studies. a full listing of the states is available in the final reports of the and studies at http://www.ii.fsu.edu/ plinternet_reports.cfm. thus, the findings below reflect information technology and libraries | june public libraries and internet access across the united states: a comparison by state – paul t. jaeger, john carlo bertot, charles r. mcclure, and miranda rodriguez paul t. jaeger (pjaeger@umd.edu) is an assistant professor at the college of information studies at the university of maryland; john carlo bertot (bertot@ci.fsu.edu) is professor and associate director of the information use management and policy institute, college of information, florida state university; charles r. mcclure (cmcclure@ci.fsu.edu) is francis eppes professor and director of the information use management and policy institute, college of information, florida state university; and miranda rodriguez (mrodrig @umd.edu) is a graduate student in the college of information studies at the university of maryland. public libraries and internet access | jaeger, bertot, mcclure, and rodriguez only those states for which both the and stud­ ies were able to provide analysis. n public libraries and the internet across the states overview of to as the public library and the internet studies have been ongoing since , the questions asked in the biennial studies have evolved along with the provision of internet access in libraries. the questions have varied between surveys, but there have been consistent questions that allow for longitudinal analysis at the national level. the study introduced the analysis of the data at both the national and the state levels. with both the and studies providing data at the state level, some longitudi­ nal analysis at the state level is now possible. overall, there were a number of areas of consistent data across the states from to . most states had fairly similar, if not identical, percentages of library outlets offering public internet access between and . for the most part, changes were increases in the percentage of library outlets offering patron access. further, the average number of hours open per week in ( . ) and in ( . ) were very similar, as were the percentages of library outlets reporting increases in hours per week, decreases in hours per week, and no changes in hours per week. while these numbers are consistent, it is not known whether this average number of hours open, or the distribution of the hours open across the week, is sufficient to meet patron needs in most communities. data across the states also indicated that physical space is the primary reason for the inability of libraries to add more workstations within the library building. there was also consistency in the findings related to upgrades and replacement schedules. changes and continuities from to while the items noted above show some areas of stability in the internet access provided by public libraries across the states, insights are possible in the areas of change for libraries overall or in the libraries that are leading in particular areas. table details the states with the highest average number of hours open per public library outlet in and . between and , the national average for the number of hours open increased slightly from . hours per week to . hours per week. this increase is reflected in the numbers for the individual states in , which are generally slightly higher than the numbers for the individual states in . for example, the top state in averaged . hours per outlet each week, while the top state in averaged . hours. the top four states—ohio, new jersey, florida, and virginia—were the same in both years, though with the top two switching positions. this demonstrates a continuing commitment in these four states by state and local government to ensure wide access to public librar­ ies. these states are also ones with large populations and state budgets, presumably fueling the commitment and facilitating the ability to keep libraries open for many hours each week. while the needs of patrons in other states are no less significant, the data indicate that states with larger populations and higher budgets, not surpris­ ingly, may be best positioned to provide the highest levels of access to public libraries for state residents. the other six states in the top ten were not in the top ten. the primary reason for this is that the six states in increased their hours more than other states. note that the fifth­ranked state in , south carolina, averaged hours per outlet each week, which is less than the tenth­ranked state in , illinois, at . hours. simply by maintaining the average number of hours open per outlet between and , south carolina fell from fifth to out of the top ten. these differ­ ences are reflected in the fact that there is nearly a ten­ hour difference from first place to tenth place in ; yet only a six­hour discrepancy exists from first place to tenth in . these numbers suggest that hours of operation may change frequently for many libraries, indicating the need for future evaluations of operational hours in rela­ tion to meeting patron demand. table displays the states with the highest average number of public access workstations per public library in and . the national averages between and also showed a slight increase from . workstations table . highest average number of hours open in public library outlets by state in and . new jersey . . ohio . . ohio . . new jersey . . florida . . florida . . virginia . . virginia . . south carolina . . indiana . . utah . . pennsylvania . . new mexico . . washington, d.c. . . rhode island . . maryland . . alabama . . connecticut . . new york . . illinois . national: . national: . in to . workstations in . a key reason for this slow growth in the number of workstations appears to have a great deal to do with limitations of physical space in libraries; in spite of increasing demands, space con­ straints often limit computer capacity. unlike table , the comparisons between and in table do not show across­the­board increases from to . in fact, florida had the highest average of workstations per library outlet in both and , but the average number decreased from . in to . in . it is interesting to note that florida has a significantly higher number of workstations than the next highest state in both and . in contrast, many of the states in the lower half of the top ten in had sub­ stantially lower average numbers of workstations in than in . in there were an average of seven more computers in spot two than spot ten; in , there were only an average of four more computers from spot two to ten. the large increases in the number of workstations in some states, like nevada, michigan, and maryland, indicate sizeable changes in budget, numbers of outlets, and/or population size. also of note is the significant drop of the average number of workstations in kentucky, declining from . in to fewer than in . a possible explanation is that, since kentucky libraries have been leaders in adopting wireless technologies (see table ), the demand for workstations has decreased as libraries have added wireless access. five states appear in the top ten of both years— florida, indiana, georgia, california, and new jersey. the average number of workstations in indiana, california, and georgia increased from to , while the aver­ age number of workstations in florida and new jersey decreased between and . some of the decreases in workstations can be accounted for by increases in the availability of wireless access in public libraries, as librar­ ies with wireless access may feel less need to add more networked computers, relying on patrons to bring their own laptops. such a strategy, of course, will not increase access for patrons who cannot afford laptops. some libraries have sought to address this issue by having lap­ tops available for loan within the library building. the states listed in table had the highest average levels of wireless connectivity in public library outlets in and . the differences between the numbers in and reveal the dramatic increases in the avail­ ability of wireless internet access in public libraries. the national average in was . percent, but in , the national average had more than doubled to . percent of public libraries offering wireless internet access. this sizeable increase is reflected in the changes in the states with the highest levels of wireless access. every position in the ratings in table shows a dra­ matic jump from to . the top position increased from percent to . percent. the tenth position increased from . percent to . percent, an increase of nearly two­and­a­half times. these increases show how much more prominent wireless internet access has become in the services that public libraries offer to their communities and to their patrons. four states appear on both the and lists— virginia, kentucky, rhode island, and new jersey. these four states all showed increases, but the rises in some table . highest average number of public access workstations in public library outlets by state in and . . florida . . florida . . kentucky . . indiana . . new jersey . . nevada . . georgia . . michigan . . utah . . maryland . . rhode island . . georgia . . indiana . . arizona . . texas . . california . . california . . new jersey . . south carolina . . virginia . new york . national: . national: . table . highest levels of public access wireless internet connectivity in public library outlets by state in and . kentucky % . virginia . % . new mexico . % . connecticut . % . new hampshire . % . indiana . % . virginia . % . rhode island . % . texas . % . kentucky . % . kansas . % . new jersey . % . new jersey . % . maryland . % . rhode island . % . illinois . % . florida . % . california . % . new york . % . massachusetts . % national: . % national: . % information technology and libraries | june public libraries and internet access | jaeger, bertot, mcclure, and rodriguez other states were significant enough to reduce kentucky from the top­ranked state in to the fifth ranked, in spite of the fact that the number of public libraries in kentucky offering wireless access increased from per­ cent to percent. in both years, a majority of the states in the top ten were located along the east coast. further, high levels of wireless access may be linked in some states to areas of high population density or the strong presence of technology­related sectors in the state, as in california and virginia. smaller states with areas of dense popula­ tions, such as connecticut, rhode island, and maryland, are also among the leaders in wireless access. tables and provide contrasting pictures regarding the number of public access internet workstations in public libraries by state in and . table shows the states with the highest percentages of libraries that consistently have fewer workstations that are needed by patrons, while table shows the states with the highest percentages of libraries that consistently have sufficient workstations to meet patron needs. of note is the fact that, unlike the preceding three tables, there appears to be no significant geographical clustering of states in tables and . nationally, the percentage of libraries that consis­ tently have insufficient workstations to meet patron needs declined from . percent in to to . percent in , a change that is within the margin of error (+/­ . percent) of the question on the survey. due to the size of the change, it is not known if the national decline was a real improvement or simply a reflection of the margin of error. washington, d.c., oregon, new mexico, idaho, and california appear on the lists for both and in table . washington, d.c. had the highest percentage of libraries reporting insufficient workstations in both years, though there was a significant decrease from percent of libraries in to percent of libraries in . in this case, the significant drop represents major strides forward to providing sufficient access to patrons in washington, d.c. similarly, though california features on both lists, the percentages dropped from . percent in to . percent in , a decline of more than half. states like these are obviously making efforts to address the need for increased workstations. overall, eight out of ten positions in table remained constant or saw a decline percentage in each position from to , indicating a national decrease in libraries with insufficient workstations. in sharp contrast, fewer than percent of nevada libraries in reported insufficient workstations, placing well out of the top ten. however, in nevada ranked second, with . percent of public libraries reporting insufficient workstations to meet patron demand. with nevada’s rapidly growing population, it appears that the demand for internet access in public libraries may not be keeping pace with the population growth. the percentage of public libraries reporting suffi­ cient workstations to consistently meet patron demands increased slightly at the national level from . percent in to . percent in , again well within the margin of error (+/­ . percent) of the question. however, in table , the top ten positions in all fea­ ture lower percentages than the same positions in . in the top­ranked state had . percent of libraries able to consistently meet patron needs for internet access, but the top­ranked state in had only percent of libraries able to consistently meet patron access needs. table . public library outlet public access workstation availability by state in and –consistently have fewer workstations than are needed . washington, d.c. % . washington, d.c. . % . california . % . nevada . % . florida % . oregon . % . new mexico . % . new mexico . % . oregon . % . tennessee . % . utah . % . alaska . % . south carolina . % . idaho % . kentucky . % . california . % . alabama . % . new york . % . idaho . % . rhode island % national: . % national: . % table . public library outlet public access workstation availability by state in and —always have a sufficient number of workstations to meet demand. . wyoming . % . louisiana % . alaska . % . new hampshire . % . kansas . % . north carolina . % . rhode island . % . arkansas . % . new hampshire . % . wyoming . % . south dakota . % . mississippi . % . georgia % . missouri . % . arkansas . % . vermont . % . vermont . % . nevada . % . virginia . % . pennsylvania . % west virginia . % national: . % national: . % � information technology and libraries | june four states—new hampshire, arkansas, wyoming, and vermont—appear on both the and lists. the national increase in the sufficiency of the num­ ber of workstations to meet patron access needs and decreases in all of the top­ranked states between and seems incongruous. this situation results, however, from a decrease in range of differences among the states from to , so that the range is compressed and the percentages are more similar among the states. further, in some states, the addition of wireless access may have served to increase the overall sufficiency of the access in libraries, possibly leveling the differences among states. nevertheless, the national average of only . percent of public libraries consistently having sufficient numbers of workstations to meet patron access needs is clearly a major problem that public libraries must work to address. comparing the data of tables and demonstrates that patron demands for internet access are being met neither evenly nor consistently across the states. nationally, the percentage of public library systems with increases in the information technology budgets from the previous year dropped dramatically from . percent in to . percent in . as can be seen in table , various national, state, and local budget crunches have significantly reduced the percentages of public library systems with increases in information technology budgets. when inflation is taken into account, a stationary information technology budget represents a net decrease in funds available in real dollar terms, so the only public libraries that are not actually having reductions in their information technology budgets are those with increases in such budgets. since internet access and the accompa­ nying hardware necessary to provide it are clearly a key aspect of information technology budgets, decreases in these budgets will have tangible impacts on the ability of public libraries to provide sufficient internet access. virtually every position on table has a decrease of percent to percent from to , with the largest decrease being from . percent in to . percent in in the second position. five states—delaware, kentucky, florida, rhode island, and south carolina—are listed for both and , though every one of these states registered a decrease from to . no drop was more dramatic than south carolina’s from . percent in to percent in . overall, though, the declining information tech­ nology budgets and continuing increases in demands for information technology access among patrons cre­ ates a very difficult situation for libraries. public libraries and the internet in along with questions that were asked on both the and public libraries and the internet studies, the sur­ vey included new questions on the study to account for social changes, alterations of the policy environment, and the maturation of internet access in public librar­ ies. several findings from the new questions on the study were noteworthy among the state data. the states listed in table had the highest percentage of public library systems with increases in total operating budget over the previous year in . nationally, . percent of public library systems had some increase in their overall budget, which includes funding for staff, physical structures, collection development, and many other costs, along with technology. at the state level, three northeastern states clearly led the way, with more than percent of library systems in maryland, delaware, and rhode island benefiting from an increase in the overall operating budget. also of note is the fact that two fairly table . highest levels of public library system overall internet information technology budget increases by state in and . florida . % . delaware % . south carolina . % . kentucky . % . rhode island . % . maryland . % . delaware . % . wyoming . % . new jersey . % . louisiana % . north carolina . % . florida % . virginia . % . rhode island . % . kentucky . % . south carolina % . new mexico . % . arkansas . % . kansas % . california . % national: . % national: . % table . highest levels of public library system total operating budget increases by state in . maryland . % . delaware % . rhode island . % . idaho . % . kentucky . % . connecticut . % . virginia . % . new hampshire . % . north carolina . % . wyoming . % national: . % public libraries and internet access | jaeger, bertot, mcclure, and rodriguez � rural and sparsely populated western states—idaho and wyoming—were among the top ten. five of the states in the top ten in highest percent­ ages of increases in operating budget in were also among the top ten in highest percentages of increases in information technology budgets in . comparing table with table reveals that delaware, kentucky, maryland, rhode island, and wyoming are on both lists. in these states, increases in information technology budgets seem to have accompanied larger increases in the overall budget. an interesting point to ponder in comparing table with table is the large discrepancy between average increases in information technology budgets ( . per­ cent) with overall budgets ( . percent) at the national level. as internet access is becoming more vital to pub­ lic libraries in the content and services they provide to patrons, it seems surprising that such a smaller portion of library systems would receive an increase in information technology budgets than in overall budgets. one growing issue with the provision of internet access in public libraries is the provision of access at suf­ ficient connection speeds. more and more internet con­ tent and services are complex and require large amounts of bandwidth, particularly content involving audio and video components. fortunately, as demonstrated in table , . percent of libraries nationally indicate that their connection speed is sufficient at all times to meet patron needs. in contrast, only . percent of public libraries nationally indicate that their connection speed is insuf­ ficient to meet patron needs at all times. georgia has the highest percentage of libraries that always have sufficient connection speed at . percent. in the case of georgia, the statewide library network is most likely a key part of ensuring the majority of libraries have sufficient access speed. many of the other states that have the highest percentages of public librar­ ies with sufficient connection speeds are located in the middle part of the country. the state with the highest percentage of libraries with insufficient connection speed to meet patron demands is virginia, with per­ cent of libraries. curiously, virginia consistently ranks in the top ten of tables – . though virginia libraries have some of the longest hours open, some of the high­ est numbers of workstations, and some of the highest levels of wireless access, they still have the highest per­ centage of libraries with insufficient connection speed. only five states had more than percent of libraries with connection speeds insufficient to meet the needs of patrons at all times. this issue is significant now in these states, as these libraries lack the necessary connec­ tion speeds. however, it will continue to escalate as an issue as content and services on the internet continue to evolve and become more complex, thus requiring greater connection speeds. comparing table with table (consistently have fewer workstations than are needed) and table (always have a sufficient number of workstations to meet demand) reveals some parallels. alabama and rhode island are among the top ten states both for connection speed being consistently insufficient to meet patron needs (table ) and consistently have fewer workstations than are needed (table ). conversely, vermont and louisiana are among the top ten states both for connection speed being sufficient to meet patron needs at all times (table ) and always have a sufficient number of workstations to meet demand (table ). table displays the two leading types of internet connection providers for public libraries and the states with the highest percentages of libraries using each. nationally, . percent of public libraries rely on an internet service provider (isp) for internet access. in the states listed in table , three­quarters or more of librar­ ies use an isp, with more than percent of libraries in kentucky and iowa using an isp. the next most common means of connection for public libraries is through a library cooperative or library network, with . percent of libraries nationally using these means. in such cases, member libraries rely on their established network to serve as the connector to the internet. the library net­ work approach seems to be most effective in geographi­ cally small states. the top three on the list being three of the smallest of the states—rhode island, delaware, and west virginia—with more than percent of libraries in each of these states connecting through a network. nationally, the remaining approximately percent of table �. highest percentages of public library outlets where public access internet service connection speed is sufficient at all times or insufficient by state in sufficient to meet patrons needs at all times insufficient to meet patron needs . georgia . % . virginia % . new hampshire . % . north carolina . % . iowa . % . alaska . % . illinois % . delaware . % . ohio . % . mississippi . % . indiana . % . missouri . % . vermont . % . rhode island . % . oklahoma . % . oregon . % . louisiana . % . connecticut . % . wisconsin . % . arkansas . % national: . % national: . % information technology and libraries | june libraries connect through a network managed by a nonlibrary entity or by other means. the highest percentages of public library sys­ tems receiving each kind of e­rate discount are presented in table . e­rate discounts are an important source of technology funding for many public libraries across the country, with more than $ , , in e­rate discounts distributed to libraries between and . nationally in , . percent of public library systems received discounts for internet connectivity, . percent for telecommunications services, and . percent for internal connection costs. mississippi and louisiana appear in the top five for each of the three types of discounts. minnesota and west virginia are each in the top five for two of the three lists. many of the states benefiting the most from e­rate funding in have large rural popu­ lations spread out over a geographically dispersed area, indicating the continuing importance of e­ rate discounts in bringing internet connections to rural public libraries. maryland and west virginia are both included in the telecommunications service column of table due to proportionally large areas of these smaller states that are rural. the importance of the telecommunications dis­ counts in certain states is obviated by the fact that more than percent of public library systems in all five states listed received such discounts. in comparison, only one state has more than percent of library systems receiv­ ing discounts for internet connectivity, while no state has percent of library systems receiving discounts for internal connection costs, with the latter reflecting the manner in which e­rate funding is calculated. in spite of the penetration of the internet into virtually every public library in the united states and the general expectations that internet access will be publicly available in every library, not all public libraries offer information technology training for patrons. nationally, . percent of public library outlets do not offer technology training. table lists the states with the highest percentages of public library outlets not offering information technol­ ogy training. six of the ten states listed are located in the southeastern part of the country. the lack of resources or adequate number of staff to provide training is a leading concern in these states. not offering patron training may be strongly linked to lacking economic resources to do so. for example, the two states with the highest percentage of public libraries not offering patron training—mississippi and louisiana—are also the two states in the top five recipients of each kind of e­rate funding listed in table . if the libraries in states like these are economically struggling just to provide internet access, it seems likely that providing accompany­ ing training might be difficult as well. a further difficulty is that there is little public or private funding available specifically for training. n discussion of issues the similarities and differences among the states indi­ cate that the evolution of public access to the internet in public libraries is not necessarily an evenly distributed phenomenon, as some states appear to be consistent lead­ ers in some areas and other states appear to consistently trail in others. while the national picture is one primarily of continued progress in the availability and quality of internet access available to library patrons, the progress is not evenly distributed among the states. libraries in different states struggle with or benefit from different issues. some public libraries are limited by state and local budgetary limitations, while other libraries are seeking alternate funding sources through grant writ­ ing and building partnerships with the corporate world. some face barriers to providing access due to their geo­ graphical location or small service population. it may also be the case that the libraries in some states do not per­ ceive that patrons desire increased access. other public libraries are able to provide high­end access as a result of having strong local leadership, sufficient state and local funding, well­developed networks and cooperatives, and a proactive state library. though the discussion of the “digital divide” has become much less frequent, the state data seem to indi­ cate that there are gaps in levels of access among libraries in different states. while every state has very successful individual libraries in terms of providing quality internet table �. highest levels of types of internet connection provider for public library outlets by state in internet service provider library cooperative or network . kentucky . % . rhode island . % . iowa . % . delaware . % . new hampshire . % . west virginia . % . vermont . % . wisconsin . % . oklahoma . % . massachusetts . % wyoming . % . minnesota . % . idaho . % . ohio . % . montana . % . georgia . % . tennessee . % . mississippi . % . alabama . % . connecticut . % national: . % national: . % public libraries and internet access | jaeger, bertot, mcclure, and rodriguez access and individual libraries that could be doing a better job, the state data indicate that library patrons in different parts of the country have variations in the levels and quality of access available to them. uniformity across all states clearly will never be feasible, though, as differ­ ent states and their patrons have different needs. for example, tables , , and all display features that indicate high­level internet access in public librar­ ies—high numbers of hours open, high numbers of public access workstations, and high levels of wireless internet access. three states—maryland, new jersey, and virginia—appear in the top ten in these three lists for . further, connecticut, florida, illinois, and indiana each appear in the top ten of two of these three lists. these states clearly are making successful efforts at the state and local levels to guarantee widespread access to public libraries and the internet access they provide. gaps in access are also evident among different regions of the country. the highest percentages of library systems with increases in total operating budgets were concentrated in states along the east coast, with seven of the states listed in table being mid­atlantic or northeastern states. in con­ trast, the highest percentages of library systems relying on e­rate funding in table were concentrated in the midwest and the southeast. further, the numbers in tables and showed far greater increases in the total operating budgets than in the information technology budgets in all regions of the country. as a result, public libraries in all parts of the united states may need to seek alternate sources of funding specifically for information technology costs. as can be seen in table , the leading states in adoption of wireless technology are concentrated in the northeast and mid­atlantic. in table , southern states, particu­ larly louisiana and mississippi, had many of the highest percentages of libraries not offering any internet training to patrons. it is important to note with data from the gulfstates, however, that the effects of hurricane katrina may have had a large impact on the results reported. one key difference in a number of states seems to be the presence of a state library actively working to coordi­ nate access issues. this particular study was not able to address such issues, but evidence indicates that the state library can play a significant role in ensuring sufficiency of internet access in public libraries in a state. maine, west virginia, and wisconsin all have state libraries that apply and distribute funds at the statewide level to ensure all public libraries, regardless of size or geography, have high­end connections to the internet. the state library of west virginia, for example, applied for e­rate funding for telecommunications costs on a statewide basis and received . percent funding in , using such funding to cover not only connection costs for public libraries, but also to provide it and network support to libraries. another example of a successful statewide effort to provide sufficient internet access can be found in maryland. in the early s, maryland public library administrators agreed to let the state library use library services and technology act (lsta) funds to build the sailor network, connecting all public libraries in the state. this network predates the e­rate program by a number of years, but having an established statewide network has helped the state library to coordinate table . highest percentages of public library systems receiving e-rate discounts by category and state in internet connectivity telecommunications services internal connection costs . louisiana . % . mississippi . % . mississippi . % . indiana . % . south carolina . % . minnesota . % . mississippi % . louisiana . % . arizona . % . minnesota . % . west virginia . % . west virginia . % . tennessee . % . maryland . % . louisiana . % national: . % national: . % national: . % table . highest levels of public library systems not offering patron information technology training services by state in . louisiana . % . mississippi . % . arkansas . % . alaska % . arizona . % . georgia . % . new hampshire . % . south carolina . % . tennessee % . idaho % national: . % information technology and libraries | june applications, funding, and services among the libraries of the state. the state budget in maryland also provides other types of funding to support the state library, the library systems, and the library outlets in providing internet access. in states such as georgia, maryland, maine, west virginia, and wisconsin, the provision of internet access in public libraries is shaped not only by library outlets and library systems, but by the state libraries as well. in these and other states, the efforts of the state library appear to be reflected in the data from this study. a final area for discussion is the degree to which librarians understand how much bandwidth is required to meet the needs of library users, how to measure actual bandwidth that is available in the library, and how to determine the degree to which that bandwidth is suf­ ficient. indeed, many providers advertise that their con­ nection speeds are “up to” a certain speed when in fact they deliver considerably less. the authors have offered an analysis of determining the quality and sufficiency of bandwidth elsewhere. suffice to say that there is consid­ erable confusion as to “how good is good enough” band­ width connection quality. these types of issues frame understandings of how connected libraries in different states are and whether those connections are sufficient to meet the needs of patrons. n future research while the experience of individual patrons in particular libraries will vary widely in terms of whether the access available is sufficient to meet their information needs, the fact that the state data indicate variations in the levels and quality of access among some states and regions of the country is worthy of note. an important area of sub­ sequent research will be to investigate these differences, determine the reasons for them, and develop strategies to alleviate these apparent gaps in access. investigating these differences requires consideration of local and situational factors that may affect access in one library but perhaps not in another. for example, one public library may have access to an internet provider that offers higher speed connectivity that is not available in another location. the range of the possible local and situational factors affecting access and services is extensive. a prelimi­ nary list of the factors that contribute to being a success­ fully networked public library is described in greater detail in the study. however, additional investigation into the degree to which these factors affect access, quality of service, and user satisfaction needs to be continued. the personal experience of the authors in working with various state library agencies suggests the need for additional research that explores relationships among those states ranked highest in areas such as connectivity and workstations with programs and services offered by the state library agencies. one state library, for example, has a specific program that works directly with individual public libraries to assist them in completing the various e­rate forms. is there a link between that state library providing such assistance and the state’s public libraries receiving more e­rate discounts per capita than other states? this is but one example where investigating the role of the state library and comparing those roles and services to the rankings may be useful. perhaps a number of “best practices” could be identified that would assist the libraries in other states in improv­ ing access and services. in terms of research methods, future research on the topics identified in this article may need to draw upon strategies other than a national survey and on­site focus groups/interviews. the study, for the first time, included site visits and interviews and produced a wealth of data that supplemented the national survey data. on­site analysis of actual connection speeds in a sample of public libraries is but one example. the degree to which survey respondents know the connec­ tion speeds at specific workstations is unclear. simply because a t­ line comes in the front door, it is not nec­ essarily the speed available at a particular workstation. other methods such as log file analysis or user­based surveys of networked services (as opposed to surveys completed by librarians) may offer insights that could augment the national survey data. other approaches such as policy analysis may also prove useful in better understanding access, connectiv­ ity, and services on a state­by­state basis. there has been no systematic description and analysis of state­based laws and regulations that affect public library internet access, connectivity, and services. the authors are aware of some states that ensure a minimum bandwidth will be provided to each public library in the state and pay for such connectivity. such is not true in other states. thus, a better understanding of how state­based policies and regulations affect access, connectivity, and services may identify strategies and policies that could be used in other states to increase or improve access, connectiv­ ity, and services. the data discussed in this article also point to many other important needs in future research. libraries in certain states that seem to be frequently ranking high in the tables indicate that certain states are better able to sustain their libraries in terms of finances and usage. however, additional factors may also be key in the differ­ ences among the states. future research needs to consider the internet access in public libraries in different states in relation to other services offered by libraries and to uses of the internet connectivity in libraries, including types of online content and services available, types of training public libraries and internet access | jaeger, bertot, mcclure, and rodriguez available, community outreach, other collection issues, staffing in relation to technology, and other factors. n conclusion internet and public computing access is almost univer­ sally available in public libraries in the united states, but there are differences in the amounts of access, the kinds of access, and sufficiency of the access available to meet patron demands. now that virtually every public library has an internet connection, provides internet access to patrons, and offers a range of public computing access, the attention of public libraries must refocus on ensuring that every library can provide sufficient internet and com­ puting access to meet patron needs. the issues to address include being open to the public a sufficient number of hours, having enough internet access workstations, hav­ ing adequate wireless access, and having sufficient speed and quality of connectivity to meet the needs of patrons. if a library is not able to provide sufficient access now, the situation will only continue to grow more difficult as the content and services on the internet continue to be more demanding of technical and bandwidth capacity. public libraries must also focus on increasing provi­ sion of internet access in light of federal, state, and local governments recently adding yet another significant level of services to public libraries by “requesting” that they provide access to and training in using numerous e­gov­ ernment services. such e­government services include social services, prescription drug plans, health care, disas­ ter support, tax filing, resource management, and many other activities. the maintenance of traditional services, the addi­ tion and expansion of public access computing and networked services, and now the addition of a range of e­government services tacitly required by federal, state, and local governments, in combination, risk stretching public library resources beyond their ability to keep up. to avoid such a situation, public libraries, library sys­ tems, and state governments must learn from the library outlets, systems, and states that are more successfully providing sufficient internet access to their patrons and their communities. among these leaders, there are likely models for success that can be identified for the benefit of other outlets, systems, and states. beyond the lessons that can be learned from the most connected, however, there are also practical and logistical issues that remain beyond the control of an individual library and sometimes the entire state, such as geographical and economic factors. ultimately, the analysis of state data offered here sug­ gests that much can be learned from one state that might assist another state in terms of improving connectivity, access, and services. while the data suggest a number of significant discrepancies among the various states, it may be that a range of best practices can be identified from those more highly ranked states that could be employed in other states to improve access, connectivity, and ser­ vices. staff at the various state library agencies may wish to discuss these findings and develop strategies that can then improve access nationwide. providing access to the internet is now as established a role for public libraries as providing access to books. patrons and communities, and now government orga­ nizations, rely on the fact that internet access will be available to everyone who needs it. while there are other points of access to the internet in some communities, such as school media centers and community technology centers, the public library is often the only public access point available in many communities. public libraries across the states must continually work to make sure the access they provide meets all of these needs. n acknowledgements the and public libraries and the internet studies were funded by the american library association and the bill & melinda gates foundation. drs. bertot, mcclure, and jaeger served as the co­principal investigators of the study. more information on these studies is available at http://www.ii.fsu.edu/plinternet/. references and notes . john carlo bertot, charles r. mcclure, and paul t. jaeger, public libraries and the internet : survey results and findings (tallahassee, fla.: information institute, ), http://www.ii.fsu .edu/plinternet_reports.cfm; john carlo bertot et al., public libraries and the internet : study results and findings (tal­ lahassee, fla.: information institute, ), http://www.ii.fsu. edu/plinternet_reports.cfm (accessed mar. , ). . bertot et al., public libraries and the internet . . john carlo bertot and charles r. mcclure, “assessing the sufficiency and quality of bandwidth for public libraries,” information technology and libraries , no. ( ): – . . john carlo bertot et al., “drafted: i want you to deliver e­government,” library journal , no. ( ): – ; john carlo bertot et al., “public access computing and internet access in public libraries: the role of pub­ lic libraries in e­government and emergency situations,” first monday , no. ( ). http://www.firstmonday .org/issues/issue _ /bertot/ (accessed mar. , ). . ibid.; paul t. jaeger et al., “the and gulf coast hurricanes: evolving roles and lessons learned for public libraries in disaster preparedness and community services,” public library quarterly (in press). . there are actually nearly , service outlets in the united states. however, the sample frame eliminated bookmobiles as information technology and libraries | june well as library outlets that the study team could neither geocode nor calculate poverty measures. additional information on the methodology is available in the study report at http://www.ii.fsu .edu/plinternet/ (accessed mar. , ). . bertot et al., public libraries and the internet . . bertot, mcclure, and jaeger, public libraries and the internet ; bertot et al., public libraries and the internet . the survey instrument is available at http://www.ii.fsu.edu/pro­ jectfiles/plinternet/plinternet_appendixa.pdf. the survey instrument is available at http://www.ii.fsu.edu/projectfiles/ plinternet/ /appendix .pdf (accessed mar. , ). . bertot et al., public libraries and the internet . . paul t. jaeger, charles r. mcclure, and john carlo bertot, “the e­rate program and libraries and library consortia, ­ : trends and issues,” information technology and libraries , no. ( ): – . . bertot, mcclure, and jaeger, public libraries and the inter- net ; bertot et al., public libraries and the internet ; john carlo bertot, charles r. mcclure, and paul t. jaeger, “public libraries struggle to meet internet demand: new study shows libraries need support to sustain online services,” american libraries , no. ( ): – . . john carlo bertot and charles r. mcclure, sailor assess- ment final report: findings and future sailor development (bal­ timore, md.: division of library development and services, ). . matt richtel and ken belson, “not always full speed ahead,” new york times, nov. , . . bertot and mcclure, “assessing the sufficiency,” – . . bertot et al., public libraries and the internet . . ibid. . bertot et al., “drafted: i want you to deliver e­govern­ ment”; bertot et al., “public access computing and internet access in public libraries”; jaeger et al., “the and gulf coast hurricanes.” . paul t. jaeger et al., “the policy implications of internet connectivity in public libraries,” government information quar- terly , no. ( ): – . information technology and libraries | june i write my final president’s column a month after the midwinter meeting in seattle. you will read it as preparations for the ala annual conference in washington, d.c. are well underway. despite that discon­ nect in time, i am confident that the level of enthusiasm will continue uninterrupted between the two events. indeed, the midwinter meeting was highly charged with positive energy and excitement. the feelings are reignited if you listen to the numerous podcasts now found on the lita blog. the lita bloggers and podcasters were omni­ present reporting on all of the meetings and recording the musings of the lita top tech trendsters. by the time you have read this you will have also, hopefully, cast your ballot for lita officers and directors after having had the opportunity to listen to brief podcast interviews with the candidates. the lita board approved the election pod­ casts at the annual conference in new orleans. thanks to the collaborative efforts of the nominating committee and the bigwig members, we have this new input into our voting decision­making. the most exciting aspects of the midwinter meeting were the face­to­face, networking opportunities that make lita so great. the lita happy hour crowd filled the six arms bar and lit it up with the wonderful lita glow badges. what was particularly gratifying to me was the number of new lita members alongside those of us who have been around longer than we care to count. the net­ working that went on there was phenomenal! the other important networking opportunity for lita members was the lita town meeting led by lita vice president mark beatty. the room was packed with eager members ready to brainstorm about what they think lita should be doing after consuming a wonderful breakfast. lita’s sponsored emerging leader, michelle boule, and mark have collated the findings and will be working with the other emerging leaders to fine­tune a direction. the podcast interview of michelle and mark is an excellent summary of what you can expect in the next year when mark is president. as stated earlier, this is my last president’s column, which means my term is winding down. using lita’s strategic plan as a guide, i have worked with many of you in lita to ensure that we have a structure in place that allows us to be more adaptable to the rapidly chang­ ing world and to make sure that lita is relevant to lita members x x and not just at conferences and lita national forum. attracting and retaining new members is critical for the health of any organization and in that vein, mark and i have used the ala emerging leaders program as a jumping off point to work with lita’s emerging leaders. the bigwig group is foment­ ing with energy and excitement as they rally bloggers and have this past year launched the podcasting initiative and the lita wiki. all of these things are making it easier for members to communicate about issues of interest in their work as well as to conduct lita business. the lita blog had over nine thousand downloads of its podcasts in the first three weeks after midwinter which confirms the desire for these types of communications! i appointed two task forces that provided recommen­ dations to the lita board at midwinter. the assessment and research task force has recommended that a perma­ nent committee be established to monitor the collection of feedback and assessment data on lita programs and services. having an established assessment process will enable the board to know how well we are accomplishing our strategic plan and to keep us on the correct course to meet membership needs. the education working group has recommended the merger of two committees, the education and regional institutes committees, into one education committee. this merged committee will develop a variety of educational opportunities including online and face­to­face sessions. we hope to have both of these committees up and going later in . happily, the feedback from the town meeting parallels the recom­ mendations of the task forces. the board will be revisit­ ing the strategic plan at the annual conference using information gathered at the town meeting. we will also be looking at what new services we should be initiating. all arrows seem to be pointing towards more educational and networking opportunities both virtual and in person. i anticipate that lita members will see some great new things happening in the next year. i have very much enjoyed the opportunity to serve as the lita president this past year. the best part has been getting to know so many lita members who have such creative ideas and who roll up their sleeves and dig in to get the work done. i am very grateful for everyone who has volunteered their time and talents to make lita such a great organization. bonnie postlethwaite (postlethwaiteb@umkc.edu) is lita president / and associate dean of libraries, university of missouri–kansas city. president’s column bonnie postlethwaite this article discusses structural, systems, and other types of bias that arise in matching new records to large data- bases. the focus is databases for bibliographic utilities, but other related database concerns will be discussed. problems of satisfying a “match” with sufficient flexibility and rigor in an environment of imperfect data are presented, and sources of unintentional variance are discussed. editor’s note: this article was submitted in honor of the fortieth anniversaries of lita and ital. s ameness is a sometime thing. libraries and other information­intensive organizations have long faced the problem of large collections of records growing incrementally. computerized records in a net­ worked environment have encouraged the recognition that duplicate records pose a serious threat to efficient information retrieval. yet what constitutes a duplicate record may be neither exact nor completely predictable. levels of discernment are required to permit matches on records that do not dif­ fer significantly and records that do. n initial definitions matching is defined as the process by which additions to a large database are screened and compared with existing database records. ideally, this process of matching ensures that duplicates are not added, nor erroneous replacements made of record pairs that are not really equivalent. oclc (online computer library center, inc.) is a non­ profit organization serving member libraries and related institutions throughout the world. it is the chief database capital of the organization, and it is “owned” in a sense by the member libraries worldwide that use and contribute to it. at this writing, it contains over seventy­three mil­ lion records. this discussion focuses chiefly on oclc’s extended worldcat (xwc), though many of the issues are common to other bibliographic databases. examples of these include the research libraries group’s research libraries information network (rlin) database, pica (a european cooperative of libraries headquartered in the netherlands), and other union catalogs. the literature will demonstrate that the problems described exist in many if not most large bibliographic databases.the database contents are representations or surrogates of the objects in shared collections. individual records in xwc are com­ plex bibliographic representations of physical or virtual objects—books, films, urls, maps, slides, and much more. each of these records consists of metadata, i.e., “structured information that describes, explains, locates, or otherwise makes it easier to retrieve, use, or manage an information resource” (appendix a). the records use an xml varia­ tion of the marc communications format. for example, a record for a book might typically contain such fields for author, title, publisher, and date, and many more in addi­ tion. the representation of any one object can be quite com­ plex, containing scores of fields and subfields. such a record may be quite brief, or several thousand characters long. the depth and richness of the records varies enormously. they may describe materials in more than languages. this is a database against which millions of searches and millions of records are processed, each month. why is matching a challenge? two records describing the same intellectual creation or work (e.g., shakespeare’s othello) can vary by physical form and other attributes. two records describing both the same work and exactly the same form can differ from each other if the records were created under different rules of record description (catalog­ ing). two records intended to describe the same object can vary unintentionally if typographical or other entry errors are present in one or both. thus sorting out significant from insignificant differences is critical. an example of the challenges of developing matching software in the metadata capture project is described elsewhere. the scope of misinformation is limited to information storage and retrieval, and specifically to comparison of incoming records to candidate matches in the database. the authors define misinformation as follows: . anything that can cause two database records, i.e., representations of different items to be mistaken as representations of the same item. these can lead to inappropriate merging or updates. . the effect of techniques or processes of search that can obscure distinctions in differing items. . any case where matching misses an appropriate match due to nonsignificant differences in two records that really represent the same item. note that disinformation (the intentional effort to mis­ represent) is not considered in scope for this discussion. the assumption is that cooperation is in the interests of all parties contributing to a shared database. we do not assume that all institutions sharing the database have the same goals. misinformation and bias in metadata processing | thornburg and oskins misinformation and bias in metadata processing: matching in large databases gail thornburg and w. michael oskins gail thornburg (thornbug@oclc.org) has taught at the university of maryland and the university of illinois, and served as an adjunct professor at kent state university, and as a senior-level software engineer at oclc. w. michael oskins (oskins@oclc.org) has worked as a developer and researcher at oclc for twenty years. information technology and libraries | june information technology and libraries | june what is bias? bias can be defined as factors in the creation or processing of database records that feed on misinformation or missing information, and skew charac­ terizations of the database records in question. context—matching and bias how are matching and bias related to each other? the growth of a database is in part a function of the matching process. if matching is not tuned correctly, the database can grow or change in nonoptimal ways. another way to look at the problem is to consider the goal of success in searching, and the need to know when to stop. human beings recognize that failure to find the best information for a given problem may be costly. finding the best information when less would suffice may also be costly. systems need to know this. for a large shared data­ base, hundreds of thousands of records may be processed in a day; the system must be as efficient as possible. what are some costs? fail to match when one should, and duplicates may proliferate in the database. match badly, and there is risk of merging multiple records that do not represent the same item. a system of matching can fail in more than one way. balance is needed. . searches, which are based on data in the incom­ ing record, may be too precise to find legitimate matches. loosen the criteria too much, and the search may return too many records to compare. . once retrieved, candidate matches are evaluated. compare candidates too narrowly, and records with insignificant differences will be rejected. fail to take note of salient differences between incom­ ing record and database record, and the match will be wrong, undetected, and potentially hard to detect in the future. the goals vary in different matching projects. for some projects, setting “holdings,” the indication that a member library owns a copy of something, is the main goal of the processing. this does not involve adding, replacing, or merging database records. for other projects, the goal is to update the database, either by replacing matched records, merging multiple duplicate records into one, or by adding new records if no match is found in the database. for the latter, bad matching could compromise database contents. n background hickey and rypka provide a good review of the problems of identifying duplicates and the implications for match­ ing software. their study notes concerns from a variety of library networks including that of the university of toronto (utlas), washington library network (wln), and research libraries group (rlin). they also refer­ ence studies on duplicate detection in the illinois state­ wide bibliographic database and at oak ridge national laboratories. background discussion of broader misinfor­ mation issues in shared library catalogs can be found in bade’s paper. a good, though dated, review of duplicate record problems can be found in the o’neill, rogers, and oskins article. the authors discuss their analysis of differences in records that are similar but not identical, and which elements caused failure to match two records for the same item. for example, when there was only one differing element in a pair, they found that element was most often publication date. their study shows the difficulties for experts to determine with certainty that a bibliographic record is for the same item. problems of typographical errors in shared biblio­ graphic records come under discussion by beall and kafadar. their study of copy cataloging errors found only . percent were corrected later by libraries, though the ordinary assumption is that copy cataloging will be updated when more information is available for an item. pollock and zamora report on a spelling error detection project at chemical abstracts service (cas) and charac­ terize the types of errors they found. chemical abstracts databases are among the most searched databases in the world. cas is usually characterized as a set of sources with considerable depth and breadth. of the four most common typographical errors they describe, errors of omission are most common, with insertion second, substitution third, and transposition fourth. over percent of the errors they found were single letter errors. this is in agreement with the findings of o’neill and aluri, though the databases were substantially different. another study on moving­ image materials focuses on problems of near­equivalents in cataloging. yee suggests that cataloging practice tends to lead to making too many separate records for near equivalents. owen gingerich provides insight in the use of holdings information in oclc and other bibliographic utilities such as rlin for scholarly research in locating early editions of copernicus’ de revolutionibus. among other sources, he used holdings information in multiple bibliographic utilities to help in collecting a census of copies of de revolutionibus, and plotting its movements through europe in the sixteenth century. his article high­ lights the importance of distinguishing very similar items for scholarly research. shedenhelm and burk discuss the introduction of vendor records into oclc’s worldcat database. their results indicate that these minimal­level records increase the duplication rate within the database and can be costly to upgrade. (see further discussion in the section change in contributor characteristics below.) one problem in analysis of sources of mismatch in previous studies is that there is no good way to detect and charac­ public libraries and internet access | jaeger, bertot, mcclure, and rodriguez misinformation and bias in metadata processing | thornburg and oskins terize typos that form real words. jasco reviews studies characterizing types and sources of errors. sheila intner compares the quality issues in the databases of oclc and the research libraries group (rlg) and finds the issues similar. intner used matched samples of records from both worldcat and rlin to list and compare types of errors in the records. she noted that while the perception at that time was that rlin had higher­quality cataloging, the differences found were not statistically significant. jeffrey beall, while focusing in his study on the full­ text online database jstor, notes the commonality of problems in metadata quality. in addition, he discusses the special quality problems in a database of scanned images. the scanning software itself may introduce typo­ graphical errors. like xwc, the database changes rapidly. o’neill and visine­goetz present a survey of quality con­ trol issues in online databases. their sections on dupli­ cate detection and on matching algorithms illustrate the commonalities of these problems in a variety of shared cataloging databases. they cite variation in title as the most common reason for failure to identify a duplicate record that should match. variations in publisher, names, and pagination were noted as common. lei zeng pres­ ents a study of chinese language records in the oclc and rlin databases. zeng discusses quality problems including ( ) format errors such as field and subfield tagging and incorrect punctuation; ( ) content errors such as missing fields and internal record inconsisten­ cies; and ( ) editing and inputting errors such as spacing and misspelling. part of her study presents the results of the prototype rule­based system developed to catch such errors. while the author refrains from comparing the quality of oclc and rlin chinese language catalog records, the discussion makes clear that the quality issues are common to a number of online databases. more work is needed on quality and accuracy of shared records in non­roman scripts, or in other lan­ guages transliterated to roman script. n types of bias to be considered specific factors that may tend to bias an attempt to match one record to another include: . violated expectations—system software expects data it does not receive, or data received is not well formed. . temporal bias—changes in rules and philosophies of record creation over time. . design bias—choices in layout of the records, which favor one type of record representation at the expense of another. . judgment calls—distinctions introduced in record representations due to differing but legitimate variation in expert judgment. oclc is a multina­ tional cooperative and there is no universal set of standards and rules for creating database records. rules of cataloging most widely used are not abso­ lutely prescriptive and are designed to allow local deviation to meet local needs. . structural bias—process and systems bias. this category reflects internal influences, inherent in the automatic processing, storage, and retrieval of large numbers of records. . growth of the database environment—whether in raw numbers of records, numbers of specific formats, numbers of foreign languages, or other characteristics that may affect efficient location and comparison of records. . changes in contributor characteristics––in the goals or focus of institutions that contribute to the database. violated expectations data may not conform to expectations. expectations about the nature of records in the data­ bases are frequently violated. what seem to be good rules for matching may not work well if the incoming data is not well formed, or simply not constructed as expected. biasing sources in the incoming data include the fol­ lowing: . typographical errors occur in titles and other parts of the record. anywhere the software has to parse text, an entry error—or even correction of an entry error by a later update—could con­ found matching. this could confound both (a) query execution and (b) candidate comparisons. basically the system expects textual data such as the name of a title or publisher to be correct, and machine­based efforts to detect errors in data are expensive to run. spelling detection techniques can compensate in some ways for data problems, but will not identify cases of real­word errors. see kukich for a survey of spelling error, real­word, and context­dependent techniques. . there is also the issue of real word differences in similar text strings. an automated system with programmed fault tolerance may wrongly equate the publisher name “mila” with “mela” when they are distinct publishers. equivalence tables can cross­reference known variations on well­known publisher names, but cannot predict merges and other organizational changes. or consider author names: are “john smith” and “jon smith” the � information technology and libraries | june � information technology and libraries | june same? this is a major problem with automated authority control where context clues may not be trustworthy. . errors of formatting of variable fields in the meta­ data contribute to false mismatch. the rules for data entry in the marc record are complex and have changed over time. erroneous placement or coding of subfields poses challenges for iden­ tification of relevant data. the software must be fault tolerant wherever possible. changes in the format of the data itself in these fields/sub­ fields may further complicate record comparisons. isbns (international standard book numbers) and lccns (library of congress control numbers) have both changed format in the recent past. . errors occur in the fields that indicate format of the information. in bibliographic records, format information is used to derive the overall type of material being described: book, url, dvd, and so on. errors in the data in combination can generate an incorrect material type for the record. . language of cataloging: this comparison has in the past caused inappropriate mismatches. the require­ ments in the new matching aimed to address this. . language in formation of queries: marc records frequently are a mixture of languages. as has been seen in other projects with intensive comparison of text, overlap in languages has the potential to confuse comparisons of short strings of text. the assumption made here is that the use of all pos­ sible syllables contained in the title should tend to mitigate language problems. nothing short of semantic analysis by the software is likely to solve such a problem, and contextual approaches to detection have had most success (in the produc­ tion environment) in carefully controlled cases. matching overall must be generic in its problem solving techniques. temporal bias large databases developed over time have their contents influenced by changes in standards for record creation, changes in contributor perception of the role of the data­ base, and changes in technology to be described. changes may include the following: . description level: e.g. changes such as book or elec­ tronic book. these have evolved from format­ to content­based descriptions that transcend format. over time, the cataloging rules for describing formats have changed. thus a format description created earlier might inadvertently “mismatch” the newer description of exactly the same item. for example, the rules for describing a book on a cd originally emphasized the cd format, whereas now, the emphasis might be shifted to focus on the intellectual content, the fact that it is a book. . the role of the database once perceived as chiefly repository or even backup source for a given library has become a shared resource with responsibilities to a community larger than any one library. . over time, the use of the database may change. (this is further discussed in the section on growth of the environment later.) searching has to satisfy the reference function of the database, but match­ ing as a process also relies on searching, and its goals are different. . varied standards worldwide challenge coopera­ tion. while u.s. libraries usually follow aacr and use the marc communications format, other parts of the world may use unimarc and country­specific cataloging rules. for instance, the pica bibliotekssystem, which hosts the dutch union catalog, used the prussian cataloging rules, which tended to focus on title entries. the switch to the rak was made by the early nineties. . some libraries may not use any form of marc but submit a spreadsheet that is then converted to marc. there is some potential for ambiguities in those conversions due to lack of : correspon­ dence of parts. . even within a country, standards change over time, so that “correct” cataloging in one decade may not match that in a later period. neither is wrong, in its own temporal context, but each results in different metadata being created to describe the same item. intner points out that oclc’s database was initi­ ated a full decade before rlg implemented rlin, and rlin started almost the same time as the aacr publication. thus rlin had many fewer pre­aacr records in its database, while worldcat had many more preexisting records to try to match with the newer aacr forms. . objects referenced in the database may change over time. for instance, a record describing an elec­ tronic resource may point to a location no longer valid for that resource. . vendor records are created as advance advertis­ ing, but there is no guarantee the records will be updated later. estimating the time before updates occur is impossible. . records themselves change over time as they are copied, derived, and migrated into other systems. they may be enhanced or corrected in any system where they reside. so when they return to the origi­ nating database, they may have been transformed so far as to be unrecognizable as representations of the same item. this problem is not unique to xwc; public libraries and internet access | jaeger, bertot, mcclure, and rodriguez �misinformation and bias in metadata processing | thornburg and oskins � it is a challenge for any shared database where export of records and reentry is likely. design bias the title, author, publisher, place of publication, and other elements of a record, designed in a time when most of the contents of a library were books, may not appear as clear or usable for other forms of informa­ tion, such as web sites or software. there is a risk to any design of a representation for an object, that it may favor distinctions in one format over another. or representations imported from other schemes may lose distinctions in the crosswalk from one scheme to another. a crosswalk is a mechanism for the mapping of data elements/content from one metadata scheme to another. dublin core and marc are just two examples of schemes used by library professionals. software exists to convert dublin core metadata to marc for­ mat, but the process of converting less complex data to a scheme of more structured data has inevitable limita­ tions. for instance, dublin core has “subject” while marc has dozens of ways to indicate subject, each with a different kind of designation for subject aspects of an item. see discussion in beall. libraries commonly exchange or purchase records from external sources to reduce the volume or costs of in­house cataloging. if an institution harvests metadata from multiple sources, there can be varying structures, content standards, and overall quality, all of which can make record compari­ sons error prone. while library and information science professionals have been creating metadata in the form of catalog records for a long time, the wider community of digital repositories may be outside the lis commu­ nity, and have varied understanding of the need for consistent representations of data. robertson discusses the challenges of metadata creation outside the library community. museums and archives may take a dif­ ferent view of what quality standards in metadata are. for example, for a museum, extensive detail about the provenance of an object is necessary. archives often record information at the collection level rather than the object level; for example, a box of miscellaneous papers, as opposed to a record for each of the papers within the box. educators need to describe resources such as learning objects. a learning object is any entity, digital or nondigital, which can be used, reused, or referenced during technology­supported learning for these objects a metadata record using the ieee lom standard may be used. while this is as complex as a marc record, it has less bibliographic description and more focus on description of the nature and use of the learning object. in short, for one type of institution the notion of appropriate granularity of description may be too detailed or too vague for the needs of another type of institution. judgment calls two persons creating independent records for the same item exercise judgment in describing what is most impor­ tant about the object. one may say it is a book with an accompanying cd, another may say it is software on a cd, accompanied by a book of documentation. another example of legitimate variation is the choice of use of ellipses […] to leave out parts of long titles in a metadata description. one record creator may list the whole title, another may list only the first part followed by the mark of ellipsis to indicate abbreviation of the lengthy title. either is correct, but may not match each other without special techniques. see appendix b for the perils of ellipsis handling. the form of name of a publisher, given other occur­ rences of a publisher name in a record, may be abbrevi­ ated. for instance, in one place the corporate author who is also the publisher might be listed in the author field as “department of health and human services” and then abbreviated—or not—in the publisher area as “the department.” note that there are limitations inherent to the valida­ tion of any system of matching, in that human reviewers may not be able to determine whether two representa­ tions in fact describe the same item. structural bias . process bias refers to any features of the software which at run­time may change the way matching is carried out, whether by shortening or lengthen­ ing the analysis, or otherwise branching the logical flow. this can arise from many sources, including but not limited to the following factors. a. there is need for efficient processing of large num­ bers of incoming records. this can force an empha­ sis on speedy matching. that is, matching not required to replace records tends to be optimized to stop searching/matching as early as is reason­ able. in the case where unique key searching finds a single match to an incoming record, it is fairly easy for the software to “justify” stopping. if there are multiple matches found, more analysis may be needed before the decision to stop matching can be made. over time the numbers of records processed has increased enormously. b. matching needs to exploit “unique” keys to speed searching, yet these may not prove to be unique. though agreements are in place for use of numeric keys such as isbns, creation of these keys is not under the control of any one organization. information technology and libraries | june information technology and libraries | june c. problems arise when brief records are com­ pared with fuller records. comparisons may be biased inadvertently towards false matches. such sparseness of data has been identified as a problem in rlin matching as well as in xwc. d. at the same time there is bias toward less generic titles in matching. requirements of sys­ tem throughput mandate an upper limit on the size of result set that the matching software will even attempt to analyze. this upper limit could tend to discriminate against effective retrieval of generic titles. matching will reject very large results sets of searches. so the query that has fewer title terms may tend to retrieve too much. titles such as “proceedings” or “bulletin” may be difficult to match if insufficient other informa­ tion is present in the record for the query to use. ironically this can mean addition of more generic titles to the database, since what is there is in effect less findable. e. transparency can contribute to bias in that, for each layer of transparency a layer of opacity may be added, when information is filtered out from a user’s view. that user may be a human or an application. openurl access to “appropriate copy” is an example from the standards world. the complexity of choosing among multiple online copies has become known as the “appro­ priate copy” problem. there are a number of instances where more than one legitimate copy of an electronic article may exist, such as mir­ roring or aggregator databases. it is essentially a problem of where and how to introduce localiza­ tion into the linking process. appropriateness reflects the user’s context, e.g., location, license agreements in place, cost, and other factors. . systems bias. what is this, really? the database can be seen as “agent.” the weight of its own mass may affect efforts to use its contents. a. for maintainers of large database systems, the goals of database repository and search engine may be somewhat at odds. yet librarians do make use of the database as reference source. b. search strategies for the software that acts as a user of the database is necessarily developed and optimized at a certain point in time. yet a river of new information flows into this data­ base. . if the numbers of types of entries in various database indexes grows nonproportion­ ally, search strategies that worked well in the past could potentially fall “out of tune” with the database contents. see growth of the environment section below. . change in proportions of languages in the database may render an application’s use of stopword lists less effective. . if changes in technology or practice result in new forms of material being described in the database, the software searches using material type as a limiter may not work properly. the software is using abstractions provided by the database, and they need to be kept synchronized. c. automated query construction presents its own problems. the use of boolean searching [term a and term b and term c] is quite restrictive in the sense that there is no “halfway” or flex for a record being included in a set of candidates. matching starts with the most specific search to avoid too­high numbers of records retrieved, and all it can do is drop or rearrange terms from a query in the effort to broaden the results. d. disconnects in metadata object creation/revision are another problem. links can point to broken uris (uniform resource identifiers). controlled vocabularies can drift or expand. even more confusing, a uri that is not broken may point to content which has changed to the point where the metadata no longer describes the item it once did. at one extreme, bruce and hillmann describe the curious case of citation of judicial opinions, for which a record of the opinion may be created as much as eighteen months before the volume with the official citation is printed, and thus the official citation cannot be created. e. expectations for creation of metadata play a role as well. traditional cataloging has generally had an expectation that most metadata is being cre­ ated once and reused. yet current practice may be more iterative, and must be, if such problems as records with broken internet uris are to be avoided. f. loss of synchronization can subvert process­ ing. note that other elements of metadata may become divorced or out of synch with the origi­ nal target /purpose. the prefix to an isbn was originally intended to describe the publisher, but is now an unreliable discriminator. numeric keys intended to identify items uniquely can retrieve multiple items, if the scheme for assign­ ing them is not applied consistently. in the worst case, meaningful data elements may become so corrupted as to be useless for record retrieval or even comparison of two records. g. ownership issues can detract from optimal data­ base management. member institutions’ percep­ tions of ownership of individual records can conflict with the goals of efficient search and retrieval. members may resist the idea of a “bet­ public libraries and internet access | jaeger, bertot, mcclure, and rodriguez misinformation and bias in metadata processing | thornburg and oskins ter” record being merged with a “lesser” one. so systems have ways of ranking records by source or contents with the general goal of trying to avoid losing information, but with the specific effect of refraining from actions that might be enriching in a given case. growth of the database environment a shared database can grow in unpredictable ways. a change in the relative proportions of different types of materials or topical coverage can render once­effective searches ineffective due to large result sets. an example of this is the number of internet­related entries in xwc. a search such as “dog” restricted to “internet­related” entries in retrieved thirty­four hits. this might be a manageable number. but in , entries were in the result set. similarly with subject headings, one search on “computer animation” retrieved fourteen hits in , and in . in both cases the result sets grew from manageable to “too large” over time. the increase in the number of foreign language entries in a database can cause problems. just determining what language an entry is in can be difficult, and records may contain multiple languages. also, such languages as chinese, japanese, and korean can overlap. chinese syllables such as: “a, an, to, no, jan, ka, jun, lung, sung, i, lo, la, le, so, sun, juan,” seen out of context might be chinese or any one of several other languages. determining appropriate handling of stopwords and other rules for effective title matching becomes more complex as more languages populate the database. changes in contributor characteristics copy cataloging practices in an institution can affect xwc indirectly. an institution previously oriented to fixing downloaded records may adopt a policy of refrain­ ing from changing downloaded records. historical inde­ pendence of libraries is one illustration. prior to the s, most libraries did not share their cataloging with other libraries. many institutions, especially smaller ones, were outside the loop and did things their own way. they used what rules they felt were useful, if they used any rules at all. later they converted sparse and poorly formed data into marc records and sent them to oclc for matching, perhaps in an effort to get back a more complete and useful record. yet the matching process is not always able to distinguish or interpret these local dialects. changes in specialization of cata­ loging staff at an institution, or cutbacks in staff can lead to reduced facility in providing original cataloging. outsourcing of cataloging work can affect handling of specialized materials as well. the introduction of vendor records and their characteristics has been noted by shedenhelm and burk. as they note, these records are very brief bibliographic records originally designed to advertise an item for sale by the vendor. these mini­ mal level records have a relatively high degree of dupli­ cation with existing records ( . percent in their study) and because of their sparseness can increase the cost of cataloging. changes in the proportion of contribu­ tors who create records in non­marc formats such as dublin core can affect the completeness of bibliographic entries. the use of such formats, meant to facilitate the entry of bibliographic materials, does come with a cost. group cataloging is a process whereby smaller libraries can join a larger set of institutions in order to reduce costs and facilitate cataloging. this larger group then contributes to oclc’s database as an entity. the growth of group cataloging has resulted in the addition of more records from smaller libraries, which may in the future have an effect on searching/matching in xwc worldcat overall. internationalization may be a factor as well. the marc format is an anglo­based format with english­language­based documentation. rapid inter­ national growth thrusts a broader range of traditions into a marc/oclc world. the role of character sets is heightened as the database grows. a cyrillic record may not be confidently matched to a transliterated record for the same item. although worldcat has a long his­ tory with cjk records, marc and worldcat are not yet accustomed to a wide repertoire of character sets. now, however, xwc is an environment in which expanding character coverage is possible, and likely. future research n we need more systematic study of the types of errors/omissions encountered in marc record cre­ ation. n how can the process of matching accomodate objects that change over time? n how does the conversion from new metadata schemes affect matching to marc records? does it help to know in what format a record arrived, or under what rules it was created? n how can we address sparseness in vendor records or legal citations? how can we deal with other advance publication issues? n how do changes in philosophy of the database affect the integrity of the matching process? n conclusions in this review we have seen that characterizing metadata at a high level is difficult. challenges for adding to a large, complex database include some of the following: information technology and libraries | june information technology and libraries | june n rules for expert creation of metadata inevitably change over time. n the object of the metadata itself may change, more often than may be convenient. n comparisons of briefer records to records that are more elaborate descriptions can have pitfalls. search and comparison strategies for such record pairs are challenged by the need to have matching algorithms that work for every scenario. n changes within the database may themselves con­ tribute to exacerbation of matching problems if duplicates are added too often, or records are merged that actually represent different contents. because of the risk, policies for merging and replacing records tend to be conservative, but this does not always favor the greatest efficiency in database processing. n changes in the membership sharing a database are likely to affect its shape and searchability. n newer schemes of metadata representation are likely to challenge existing algorithms for determining matches. references . national information standards organization, under- standing metadata (bethesda, md.: niso pr., ), . http:// www.niso.org/standards/resources/understanding metadata. pdf (accessed feb. , ). . library of congress, “marc concise format for bibliographic data ( ).” http://www.loc.gov/marc/ bibliographic/ecbdhome.html (accessed nov. , ). . gail thornburg, “matching: discrimination, misinforma­ tion, and sudden death,” informing science conference, flag­ staff, ariz., june . . thomas b. hickey and david j. rypka, “automatic detec­ tion of duplicate monographic records,” journal of library auto- mation , no. (june ): – . . david bade, “the creation and persistence of misinfor­ mation in shared library catalogs,” occasional paper no. , (graduate school of library and information science, univer­ sity of illinois at urbana–champaign, apr. ). . edward t. o’neill, sally a. rogers, and w. michael oskins, “characteristics of duplicate records in oclc’s online union catalog,” library resources and technical services , no. ( ): – . . jeffrey beal and karen kafadar, “the effectiveness of copy cataloging at eliminating typographical errors in shared bibliographic records,” library resources & technical services , no. (apr. ): – . . j. j. pollock and a. zamora, “collection and characteriza­ tion of spelling errors in scientific and scholarly text,” journal of the american society for information science , no. ( ): – . . edward t. o’neill and rao aluri, “a method for cor­ recting typographical errors in subject headings in oclc records,” research report # oclc/opr/rr­ / ( ). . martha m. yee, “manifestations and year­equivalents: theory, with special attention to moving­image materials,” library resources and technical services , no. ( ): – . . owen gingerich, “researching the book nobody read: the de revolutionibus of nicolaus copernicus,” the papers of the bibliographical society of america , no. ( ): – . . laura d. shedenhelm and bartley a. burk, “book vendor records in the oclc database: boon or bane?” library resources and technical services , no. ( ): – . . peter jasco, “content evaluation of databases,” in annual review of information science and technology, vol. (medford, n.j.: information today, inc., for the american society for infor­ mation science, ), – . . sheila intner, “quality in bibliographic databases: an analysis of member­controlled cataloging of oclc and rlin,” advances in library administration and organization ( ): – . . jeffrey beall, “metadata and data quality problems in the digital library,” journal of digital information , no. ( ): – . . edward t. o’neill and diane vizine­goetz, “quality control in online databases,” annual review of information sci- ence and technology (washington, d.c.: american society for information science, ). . lei zeng, “quality control of chinese­language records using a rule­based data validation system. part : an evalua­ tion of the quality of chinese­language records in the oclc oluc database,” cataloging and classification quarterly , no. ( ): – . lei zeng, “quality control of chinese­language records using a rule­based data validation system. part : a study of a rule­based data validation system for online chinese cata­ loging,” cataloging and classification quarterly , no. ( ): – . . anglo-american cataloguing rules, nd ed., rev. (chi­ cago: ala, ). . karen kukich, “techniques for automatically correct­ ing words in text,” acm computing surveys , no. ( ): – . . gail thornburg, “the syllables in the haystack: techni­ cal challenges of non­chinese in a wade­giles to pinyin con­ version,” information technology and libraries , no. ( ): – . . hartmut walravens, “serials cataloguing in germany: the historical development,” cataloging and classification quar- terly , no. / ( ): – ; instruktionen für die alphabetischen kataloge der preuszischen bibliotheken vom . mai . ausg. in der fassung vom . august (berlin: behrend & co., ). . richard greene, e­mail message to author, nov. , ; regeln für die alphabetische katalogisierung: rak / irmgard bou­ vier (wiesbaden, germany: l. reichert, , c ). . intner, “quality in bibliographic databases.” . richard greene, e­mail message to author, feb. , . . beall, “metadata and data quality problems in the digital library.” . r. john robertson, “metadata quality: implications for library and information science professionals,” library review , no. ( ): – . public libraries and internet access | jaeger, bertot, mcclure, and rodriguez misinformation and bias in metadata processing | thornburg and oskins . ieee. learning technology standards committee, “wg : learning objects metadata.” http://ltsc.ieee.org/wg (accessed feb. , ). . ibid. . orien beit­arie et al., “linking to the appropriate copy: report of a doi­based prototype,” d-lib , no. (sept. ). . thomas r. bruce and diane i. hillmann,“the continuum of metadata quality: defining, expressing, exploiting,” in meta- data in practice (chicago: ala, ), – . . shedenhelm and burk, “book vendor records in the oclc database.” information technology and libraries | june information technology and libraries | june appendix a. sample cdfrecord record from the xwc database cgm a vf bcahru mr baaafu r fr mleng dlcamim dlc lp u.s. copyright office xxu mr vbe ­ (viewing copy) fgb ­ (ref print) fpa ­ (master­ pos) othello (motion picture : welles) the tragedy of othello­­ the moor of venice / a mercury production, [films marceau?] ; directed, produced, and written by orson welles. u.s. ; [morocco?] france :films marceau, ; [morocco?: :s.n., ?] ;united states : united artists, . videocassettes of (ca. min.) :sd., b&w ; / in. viewing copy. reels of on (ca. ft.) :sd., b&w ; mm. ref print. reels of on (ca. ft.) :sd., b&w ; mm. masterpos. copyright: orson welles; sep ; lp . reference sources cited below and m/b/rs preliminary cataloging card list title as othello. photography, anchisi brizzi, g.r. aldo, george fanto ; film editors, john shepridge, jean sacha, renzo lucidi, william morton ; music, francesco lavagnino, alberto barberis. orson welles, suzanne cloutier, micheaÌ l macliamoÌ ir, robert coote. director, producer, and writer credits taken from focus on orson welles, p. . lc has u.s. reissue copy.dlc new york times, / / . an adaptation of the play by william shakespeare. reference sources used: new york times, / / ; international motion pic­ ture almanac, , p. ; focus on orson welles, p. ­ ; monthly film bulletin, v. , no. , p. ; index de la cineÌ matog­ raphie française, , p. . received: / / from lc video lab;viewing copy; preservation, made from ref print, paperwork in acq: copyright­­material movement form file, lwo ; copyright collection. received: / / ; ref print;copyright deposit; copyright collection. received: / ; masterpos;gift; afi theatre collection. othello (fictitious charac­ ter)drama. public libraries and internet access | jaeger, bertot, mcclure, and rodriguez misinformation and bias in metadata processing | thornburg and oskins plays. mim features. mim welles, orson, ­direction, production,writing, cast. cloutier, suzanne, ­cast. mac liammoÌ ir, micheaÌ l, ­ ,cast. coote, robert, ­ ,cast. copyright collection (library of congress)dlc afi theatre collection (library of congress)dlc othello. appendix b. the perils of judging near matches a. challenges of handling ellipses in titles thought to be similar incoming title: general explanation of tax legislation enacted in ... / prepared by the staff of the joint committee on taxation match: general explanation of tax legislation enacted in the th congress prepared by the staff of the joint committee on taxation incoming title: general explanation of tax legislation enacted in ... / prepared by the staff of the joint committee on taxation match: general explanation of tax legislation enacted in the th congress prepared by the staff of the joint committee on taxation incoming title: general explanation of tax legislation enacted in ... / prepared by the staff of the joint committee on taxation match: general explanation of tax legislation enacted in the th congress prepared by the staff of the joint committee on taxation incoming title: general explanation of tax legislation enacted in ... / prepared by the staff of the joint committee on taxation match: general explanation of tax legislation enacted in the th congress prepared by the staff of the joint committee on taxation b. partial matches in names which might represent the same publisher publisher comparison is challenging in an environment where organziations are regularly merged or acquired by other organziations. there is no real authority control for publishers that would help cataloguers decide on a preferred form. when governmental organizations are added to the mix, the challenges increase. below are some examples of non­match­ ing text of publisher names in records, which might or might not considered the same by a human expert. (the publisher names have been normalized.) information technology and libraries | june information technology and libraries | june . publisher name may be partially or differently recorded in two records incoming publisher: konzeptstudien kantonale planungsgruppe match: kantonale planungsgruppe konzeptstudien (word order different) incoming publisher: institut francais proche orient match: institut francais darcheologie proche orient incoming publisher: u s dept of commerce national oceanic and atmospheric administration national environ­ mental satellite data and information service match: national oceanic and atmospheric administration . publisher name may have changed due to acquisition by another organization incoming publisher: pearson prentice hall match: prentice hall incoming publisher: uxl match: uxl thomson gale incoming publisher: thomson arco match: arco thomson learning . one record may show “publisher” which is actually government distributing agency or clearinghouse such as the u.s. government printing office or national technical information service (ntis), while the candidate match shows the actual government agency. these can be almost impossible to evaluate. incoming publisher: u s congressional service match: supt g p o (here the distributor is the government printing office, listed as the publisher) incoming publisher: u s dept of commerce national oceanic and atmospheric administration national environmental satellite data and information service match: national oceanic and atmospheric administration incoming publisher: u s gpo match: u s fish and wildlife service . the publisher in a record may start with or end with the publisher in the second record. should it be called a match? good: incoming publisher trotta match: editorial trotta incoming publisher wiley match: john wiley questionable? incoming publisher prentice hall match: prentice hall regents canada incoming publisher geuthner match: orientaliste geuthner incoming publisher oxford match: distributed royal affairs oxford incoming publisher: pan union general secretariat organization states match: social science section cultural affairs pan union the goal of this paper is to describe a design—includ- ing the hardware, software, and configuration––for an open source wireless network. the network designed will require authentication. while care will be taken to keep the authentication exchange secure, the network will oth- erwise transmit data without encryption. w ireless networks are an essential tool for provid­ ing service for colleges and libraries. this paper will explain the setup of a wireless network using open­source software and inexpensive commodity hardware. open­source software was employed exclu­ sively. this allowed for flexibility in design and reduction in expense while also providing a platform for students to learn more about the internal workings of the system by examining particular sections of code in which they have interest. standard commodity hardware was used as a means of saving cost. this should allow others to repeat this design with a minimum of funding. the purpose of a network, like any resource, is to provide a service for those who own it; in this case, the patrons of a library, or students, faculty, and staff at a col­ lege. to ensure that this network serves its owners, users will be required to authenticate before gaining access. once authenticated, the central captive portal can pro­ vide different levels of service for specific user groups, including guest access, if desired. for this system, ease of access for users was the primary concern; other than using the secure socket layer for authentication, the remainder of the traffic was unencrypted. other than the base nodes, the remaining access points were connected to each other using a wireless connection in order to avoid physically connecting all access points across campus and to further reduce the expense for the deployment of the network. this was accomplished using the wds (wireless distributed system) feature on the wireless routers. all access points connect to a centralized set of servers that provide: dhcp, web­caching proxy, dns caching, radius, web server, a captive portal, and logging of network traffic. n hardware requirements for the network were relatively modest, using inexpensive wireless routers along with several linux servers built upon older pentium desktop systems. linksys wrt gs routers were chosen as the access points as they are inexpensive, readily available, and possess the ability to run custom open­source firmware. other access points could be used; however, the configuration sugges­ tions are specific to the wrt gs and may not apply to other hardware. the routing functions of the wrt gs were not used in this implementation. the servers need not be anything special; older hardware will work just fine. for this implementation, decommissioned mhz units with mb of ram and gb hard drives were used. n wireless router software in order to provide the functionality required, the units had their firmware flashed with an open­source, linux­ based operating system available from openwrt for the linksys routers (http://www.openwrt.org). support is also available for other wireless devices. “the firmware from openwrt provides a fully writable file system with pack­ age management. this allows developers the freedom to customize the devices by choosing only the packages and software that are necessary for their applications.” as the routers have limited storage, being able to hand select only the necessary components is a definite advantage. n server software for the operating system on the servers, fedora core was chosen. fedora provides the yellow dog updater, modified (yum), which eases the updating of all pack­ ages installed on the system, including kernel updates. this aids security by providing a platform for easily and frequently updating the system. fedora core is an open­ source distribution that is available for free. fedora core also comes with many other open­source packages that were used in this design, such as the apache web server. while the designers had more familiarity with fedora, other distributions are also available that provide simi­ lar benefits (suse, ubuntu, openbsd, debian, etc.). the server was run in command line mode with no graphical user interface in order to reduce the load on the server and save space on the hard drive. n captive portal in order to require authentication before gaining access to the network, a captive portal was used. some of the open source wifi hotspot implementation | sondag and feher open source wifi hotspot implementation tyler sondag and jim feher jim feher (jdfeher@mckendree.edu) is an associate professor of computer science at mckendree college in lebanon, illinois. tyler sondag (tnsondag@mckendree.edu), is a senior in computer science at mckendree college. information technology and libraries | june information technology and libraries | june desired features in the choice of the captive portal were: encrypted authentication, traffic logging, and the ability to provide different levels of service for different user groups. logging traffic allows the system administrators to identify accounts that have been misusing the network. those who inadvertently misuse the system or perhaps have had their accounts compromised can have their access temporarily disabled until they can be contacted with instructions concerning acceptable use of the net­ work. as the network must be shared by all, those who habitually abuse the resource can have their accounts per­ manently disabled. the captive portal should also redi­ rect web traffic to a login page that is served on the secure socket layer until the user logs in. chillispot was chosen as it possesses all of the features mentioned above. n server layout as can be seen in appendix a, three servers were used for this implementation. the first server was used as the main router to the internet. the second server ran a squid web caching server. it also ran a dns cach­ ing server and the freeradius server. the third was used for the captive portal. three servers were used for various reasons. first, this distributed the load. second, portions of the network that were not behind the cap­ tive portal could more easily use the services on the second server running squid, dns, and freeradius. it should be noted that three independent servers are not required; many of the services could be consolidated on two or even one single server to reduce the hardware requirements. the implementation depends upon the specific needs for the network. n server installation installing the operating system (fedora core) on each server is a relatively straightforward procedure. each machine was partitioned with mbs of swap space with the rest of the drive being an ext partition with the mount point “/”. only the minimal set of packages required were installed at this time. the first server, server # (router), was given three network interfaces, one for the internet connection, one to connect to a switch that then connects to server # (web/dns caching and radius) as well as other machines that do not connect through the captive portal, and one connecting to server # (captive portal machine). the second server, server # , only needs one interface, but the third, server # , requires two interfaces, one for the master wireless access point, and one to connect to the switch connecting this machine to the rest of the network (appendix a). ssh login for root was also disabled at this time for added security. n server # configuration for server # , very little setup was required. since this server works mainly as a router, the only major items that went into its configuration were the iptables rules, which are shown and described in appendix b. rules were set up to: n set up network address translation; n allow traffic to flow within the network; n log the traffic from the wireless portion of the net­ work; n allow for the transparent setup of the web proxy server; and n set up port knocking before allowing users to log into the router via ssh. a reference to this script was placed in the /etc/rc.d/ rc.local file so that it would run when the server boots. last was the setup of the three network interfaces in the machine. this can be done during system installation or afterwards on the fedora core based server by editing the configuration files in the /etc/sysconfig/networking­ scripts/ directory. one of the configuration files used in this implementation can be seen in appendix c. of course the configuration will change as the topology of the net­ work changes. n server # configuration the second server required significantly more setup to configure all of the necessary services that it runs. the first service added for this implementation was the web­ caching proxy server, squid. squid’s default configura­ tion file (/etc/squid.conf) is quite large; fortunately it requires little modification to get a simple server up and running. the changes made for this implementation can be seen in appendix d. the most important lines in this configuration are the last few, which enable it to act as a transparent proxy server, making it invisible to the users and requiring no setup of their browsers. as there was no need for an authoritative dns server, just dns caching for the network, dnsmasq, which is easy to configure and can handle both dhcp services as well as dns caching, was chosen. in this instance, the captive portal was used to provide dhcp services for the wireless clients; however dnsmasq was used for dynamic clients on the remaining portion of the network. dnsmasq public libraries and internet access | jaeger, bertot, mcclure, and rodriguez open source wifi hotspot implementation | sondag and feher is relatively easy to configure, requiring only one change in its default configuration file, which points to the file in which the dns server addresses are stored, in this case /etc/dnsmasq_resolv.conf. next is the configuration of freeradius server. there are two files that need to be modified for the radius server; both are in the /etc/raddb/ directory. the first is clients.conf (appendix e). in this file at least two clients must be listed, one for localhost (this machine) and one for the captive portal machine. for each machine, a pass­ word must be specified as well as the hostname for that machine. this establishes the shared key that is used to encrypt communication between the captive portal and the radius server. the second is the users file (appendix f). in this file, each user for the captive portal system must be listed with his/her password. this implementa­ tion also included a class, a session timeout (dhcp lease time), idle timeout, accounting interim interval, and the maximum upload and download speeds. if guest access is required, one or several guest accounts should be added to this file along with entries for the registered users. an entry was added for each access point so that they can obtain an ip address from the dhcp server. finally for this machine, the interface configuration file was changed according to the network specifications. for this machine the configuration is simple since it only has one interface, and the only requirement for its address is that it be on the same network as the interface on the main router server to which it is connected. n server # configuration the third server required the installation of the captive portal software, in this case chillispot. in order to install chillispot, if fedora was used for the base system, it may be possible to install it as a prepackaged binary in the form an rpm package manager (rpm) file. otherwise, if you find that you need to compile chillispot from source code, you may need to deviate from a minimal installa­ tion of the operating system and base components and also include the gnu compiler collection (gcc). when installing from source code, first download the code from the chillispot web site. once the code is down­ loaded, unzipped and untarred, installing the chillispot daemon is done by entering the directory containing the source files and entering the standard commands: ./configure make make install when chillispot is on the system, either by compiling from source or through an rpm file, two more files must be configured and copied to the proper directory, the main configuration file and the login file. the configuration file, chilli.conf, is located in the directory that contains the source files. move this file to the /etc/ directory and make the necessary changes. in this implementation, the file required several changes (appendix g). one of the more significant alterations was to change the default network range of . . . / , which would be limited to less than addresses. the address range was for the dhcp server was also expanded to allow for more users. the lower portion of the network range was left to make room for addresses that could be assigned to the wireless access points. an entry was added to allow the access points to obtain a static ip address in that lower range. after this, settings must be changed for the dns addresses given out to clients, and the address of the radius server. there is also a setting in the chillispot configuration file that allows users to access a certain list of domains without logging in. for this implementation, the decision was to allow the users access to the campus network, as well as to the dns server. next, the “radi­ ussecret” must be set. this is the same password that was entered into the clients.conf file on the radius server for this machine. it is also necessary to set the address of the page to which users will be directed. two lines must also be added to allow authentication using the physical or media access control (mac) address for the access points. all of the access points shared a common password. chillispot passes the physical address of the access point to the radius server along with this password. a separate entry must exist in the radius configuration file for each ip/physical address combination. for this setup, the redirect page was placed on this server, therefore apache (using yum) was also installed, and this server’s address was added as the web address for the redirect page (also note that the https module may be required for apache if it does not automatically install). rather than write a new page at this time, the sample page (hotspotlogin.cgi) from the chillispot source folder was copied and modified slightly (appendix h). in addi­ tion, a secure socket layer (ssl) certificate was installed on this server. this is not necessary, but it helps to avoid the warnings that pop up when a client attempts to access the login page with a browser. a few iptables rules need to be added. the first com­ mand needs to be executed in order to utilize network address translation (nat) and have the server forward packets to the outside network. /sbin/iptables ­t nat ­a postrouting ­o eth \ ­j masquerade the next is used to drop all outbound traffic originating from the access points. this prevents anyone spoofing the physical address of the access point from accessing � information technology and libraries | june � information technology and libraries | june the internet, while still allowing the access points and the chillispot server to communicate for configuration and monitoring. /sbin/iptables ­a forward ­s . . . / \ ­j drop these commands need to be executed when the chillispot machine boots, so they were placed into the /etc/rc.d/rc.local file. it may also be necessary to ensure that the machine can forward network traffic. this can be accomplished with the following command, which is also found as the first executable command from the script in appendix b: echo “ ” > /proc/sys/net/ipv /ip_forward finally, the configuration files for the interfaces were set up. n openwrt installation and configuration several ways exist to replace the default linksys firmware with the openwrt firmware. the tftp protocol can be used with both windows and linux, and one such method can be found in appendix i. in addition, other methods for using the standard web interface can be found on the openwrt web site. there are several versions of the openwrt firmware available; the newest version that uses the squashfs filesystem was chosen because it utilizes com­ pression that frees more space on the access point. openwrt comes with a default web interface that can be used for configuration, however, ssh was enabled and a script using the nvram command was used to configure each access point (see appendix j). before ssh can be used, you must telnet into the router and change the default password (which for linksys routers is ‘admin’). note: even if you decide to use the web interface, you should still change the default password. as several services that were installed with the default configuration were not used in the implementa­ tion, they were disabled once the firmware was flashed by removing the modules that boot at startup: the web interface, dnsmasq, and the firewall. this is done by deleting their entries in the /etc/init.d directory. changes were needed to set the mode of the access point, to turn on and configure the clients needing to use wds, to set the network information for the access point and then to save these settings. all of the wireless access points that communicate with each other via a wireless connec­ tion must have their physical addresses entered using a nvram command. for example, the command used for the main access point for the library would be: nvram set w _wds=”mac_ _lib mac_ _lib ” all of this is detailed in appendix j. a final set of com­ mands, which were needed for the wrt gs, are included to allow the access point to obtain its ip address from the dhcp server. these commands may not be necessary depending upon the type of access point used. since extra wireless access points are available, if an access point fails or is having problems for some reason, it is simply a matter of running a script similar to the one found in the appendix on one of the extra routers and swapping it out. n security unfortunately this system is not very secure. only the login credentials are encrypted via ssl. general data packets are in no way encrypted, so any information being transmitted is available to anyone sniffing the channel. wep and wpa could be used for encryption, but they have known vulnerabilities. other methods exist for securing the network such as wpa with radius or the use of a virtual private network, however the client setup for such systems may not be considered trivial for the typical user. therefore it was decided that it was better to inform the users that the data was not being encrypted and let them act accordingly, rather than use encryption with known flaws or invest the time required to train the general population on how to configure their mobile units to use a more secure form of encryption. as the main goal of this particular network was connectivity and not security, it was felt that this was a fair trade­ off. as new standards for wireless communication are developed and commodity hardware that supports them becomes available, this may change so that encrypted channels can be employed more easily. n conclusion this implementation is in no way completed. it is a work in progress, with many goals still in mind. also, as new features are desired, parts of the system will change to accommodate these requirements. current plans for the future are first to develop scripts to check the status of the access points and display this information to a web page. these scripts will also notify network administrators when access points go offline. this will help the adminis­ trators in making sure the system is up at all times. after this, scripts will be developed to parse the log files to find abusive activity (spamming, viruses, etc). however, the current project as described is complete and has already functioned successfully for nearly a year providing con­ nectivity for the library and portions of the mckendree college campus. public libraries and internet access | jaeger, bertot, mcclure, and rodriguez �open source wifi hotspot implementation | sondag and feher � references and notes . openwrt, wireless freedom. www.openwrt.org (accessed june , ). . the fedora project. www.fedora.redhat.com (accessed nov. , ). . yum: yellow dog updater, modified. www.linux.duke. edu/projects/yum (accessed july ). . chillispot—open source wireless lan access point controller. www.chillispot.org (accessed june , ). . squid web proxy cache. www.squid­cache.org (accessed june , ). . freeradius—building the perfect radius server. www. freeradius.org (accessed june , ). . netfilter/iptables project homepage—the netfilter.org project. www.netfilter.org (accessed aug. , ). . thomas eastep, “port knocking and other uses of ‘recent match.’” www.shorewall.net/portknocking.html (accessed aug. , ). . squid web proxy cache, “squid frequently asked questions: interception caching/proxying.” www.squid­cache. org/doc/faq/faq­ .html (accessed aug. , ). . dnsmasq—a dns forwarder for nat firewalls. www. thekelleys.org.uk/dnsmasq/doc.html (accessed june , ). . linksys.com. www.linksys.com (accessed dec. , ). . openwrtdocs/installing/tftp—openwrt. wiki.open­ wrt.org/openwrtdocs/installing/tftp?action=show&redirect =openwrtviatfp (accessed aug. , ). . openwrtdocs/installing—openwrt. wiki.openwrt.org/ openwrtdocs/installing (accessed aug. , ). appendix a. network configuration information technology and libraries | june information technology and libraries | june appendix b. iptables script—server # # this particular bit must be set to one to allow the # network to forward packets echo “ ” > /proc/sys/net/ipv /ip_forward # set up path to the internal network from internet if the # internal network initiated the connection iptables ­a forward ­i eth ­o eth ­d . . . \ ­m state ­­state established,related ­j accept # same for the chillispot subnet iptables ­a forward ­i eth ­o eth ­d . . . \ ­m state ­­state established,related ­j accept # allow the internal subnets to communicate with one another iptables ­a forward ­i eth ­d . . . ­o eth \ ­j accept iptables ­a forward ­i eth ­d . . . ­o eth \ ­j accept # allow subnet containing server to reach the internet iptables ­a forward ­i eth ­o eth ­j accept # chillispot – accept and forward packets iptables ­a forward ­i eth ­s . . . ­j accept # set up transparent proxy for wireless network, but allow # connections that go through to the campus network # to bypass proxy iptables ­t nat ­a prerouting ­i eth ! \ ­d . . . / ­p tcp ­­dport ­s . . . / \ ­j dnat ­­to­destination . . . : # nat iptables ­t nat ­a postrouting ­o eth \ ­j masquerade # simple port knocking to allow port connection adapted # from www.shorewall.net/portknocking.html another # excellent document can be found at # www.debian-administration.org/articles/ # once connection started let it continue iptables ­a input ­m state ­­state \ established,related ­j accept # if name ssh has been set, then allow connection iptables ­a input ­p tcp ­­dport ­m recent \ ­­rcheck ­­name ssh ­j accept # surround the port that opens ssh so that a sequential port # scanners will end up closing it right after opening it. iptables ­a input ­p tcp ­­dport ­m recent \ –­name ssh ­­remove ­j drop iptables ­a input ­p tcp ­­dport ­m recent \ ­­name ssh ­­set ­j drop iptables ­a input ­p tcp ­­dport ­m recent \ ­­name ssh ­­remove ­j drop # drop all packets that do not match a rule above by default iptables ­a input ­j drop appendix c. server configuration for first network card (ethernet ) # /etc/sysconfing/networking­scripts/ifcfg­eth ­ # server # # device=eth bootproto=static broadcast= . . . hwaddr= : : : : : ipaddr= . . . netmask= . . . network= . . . onboot=yes type=ethernet appendix d. /etc/squid.conf—server # #default squid port http_port # settings changed to specify memory for squid cache_mem mb cachedir ufs /var/spool/squid # allow assess to squid for all within our network acl all src . . . / . . . http_access allow all http_reply_access allow all # internal host with no externally known name so we put # our internal host name visible_hostname hostname # specifications needed for transparent proxy httpd_accel_port httpd_accel_host virtual httpd_accel_with_proxy on httpd_accel_uses_host_header on public libraries and internet access | jaeger, bertot, mcclure, and rodriguez open source wifi hotspot implementation | sondag and feher appendix e. /etc/raddb/clients.conf— server # client . . . { secret = password shortname = localhost nastype = other } client . . . { secret = password shortname = other machine } appendix f. /etc/raddb/users—server # # example of an entry for a user joeuser auth­type:=local, user­password==”passwd” class = , session­timeout = , idle­timeout = , acct­interim­interval = , wispr­bandwidth­max­up = , wispr­bandwidth­max­down = # example of an entry for an access point # the physical/mac address listed below is for the # lan side of the router/access point mac_address auth­type := local, user­password == “password” framed­ip­address = . . . , acct­interim­interval = , session­timeout = , idle­timeout = appendix g. /etc/chilli.conf—server # # used to expand the network net . . . / # used to expand the number of hosts that can connect # while still leaving a portion of the network for # infrastructure dynip . . . / # used to give static addresses to the access points statip . . . / # internal dns followed by external dns dns . . . dns . . . # radius server for the network radiusserver . . . radiusserver . . . # radius secret used radiussecret password # interface chillispot server to listens to dhcp requests dhcpif eth # specified default login page uamserver https:// . . . /cgi­bin/hotspotlogin.cgi # addresses that users can visit without authenticating uamallowed . . . , . . . , . . . / # this allows the access points to authenticate based on # mac address only, this is required to log into the access # points from the captive portal server macauth # this password corresponds with the password from the # radius users file macpasswd password information technology and libraries | june information technology and libraries | june appendix h. redirection page appendix i. method for flashing firmware of linksys router the firmware can be flashed using the built­in web inter­ face or via tftp. while help is available online for this, the procedure outlined here may also be helpful. on newer versions of the linksys routers, an older version of the linksys firmware must be installed first that supports a bug in the ping function on the router. once the older version is installed, you can exploit a bug in the ping com­ mand on the router to enable “boot wait,” which enables the router to accept a connection to flash its firmware as it is booting. detailed instructions for this installation are as fol­ lows: n first, download an old version of a linksys firmware that supports the ping bug to enable boot wait. one is available at: ftp://ftp.linksys.com/pub/network/ wrt gs_ . . _us_code.zip n download and unzip this file. n plug an ethernet patch cable into link # on the router (not the wan port) and the interface on your machine. set the ip address of your computer to a static ip address in the . . .x range, not . . . , which is used by the router. n log into router by opening a browser window and putting . . . into the address bar. (note: this is only for factory preset routers.) username: (leave blank) password: admin n click on "administration". n click on "firmware upgrade". n click "browse" and locate the old linksys firmware on your machine. n click "upgrade". n wait patiently while it flashes the firmware…. n click "setup". n click "basic setup". public libraries and internet access | jaeger, bertot, mcclure, and rodriguez open source wifi hotspot implementation | sondag and feher n choose "static ip" from the first box. n for the ip address put in " . . . ". n for the netmask put in " . . . ". n for the gateway put in " . . . ". n you can leave everything else as their default set­ tings. n choose save settings at the bottom of the page. n click on "administration". n click on "diagnostics". n click on "ping". in the “address” box put the following commands in one at a time and click on “ping”; if you see the message that the host was unreachable you have done something wrong. ;cp${ifs}*/*/nvram${ifs}/tmp/n ;*/n${ifs}set${ifs}boot_wait=on ;*/n${ifs}commit ;*/n${ifs}show>tmp/ping.log n after the last command you will see a list of all the nvram settings on the router, make sure that the line for "boot_wait" is set to on n unplug the router (the linksys router will only look for new firmware on boot). n use tftp on your linux or windows machine. n if the openwrt ­wrt gs­squashfs.bin file is not in this directory, copy the file to this directory n run the following commands at the prompt (below are the linux commands) tftp . . . tftp> binary tftp> rexmt tftp> timeout tftp> trace tftp> put openwrt­xxx­x.x­xxx.bin n the router will now reboot (it may take a very long time), when it is done rebooting, the dmz light will turn off the new firmware is now loaded onto the router. appendix j. nvram script for wireless routers ## server information stored as comments ## . . . mainap : : : : : ## . . . cl a : : : : : ## . . . lib : : : : : ## . . . lib : : : : : ## . . . lib : : : : : ## . . . car : : : : : ## same for all nvram set wl _mode=ap nvram set wl _ssid=mck_wireless nvram set wl _channel= nvram set lan_proto=dhcp ## sample configuration for a few access points. ## uncomment and run for the appropriate node. ## make sure to ## add a line for every access point you have. ## unique for lib ## allow connections to/from lib , and lib #nvram set wl _wds=” : : : : : : : : : : ” ## unique for lib ## allow connections to/from lib #nvram set wl _wds=” : : : : : ” ## unique for lib ## allow connections to/from lib #nvram set wl _wds=” : : : : : ” ## same for all nvram commit ## same for all ## this needed to be done to allow each wrt gs router ## to accept an ip address from a dhcp server. this is ## only for the wrt gs. other access point/routers ## may require something different. # cd /etc/init.d # rm s nvram # cp /rom/etc/init.d/s nvram . # vi s nvram ## place a # in front of (comment out) ## nvram set lan_proto=”static” references . thomas eastep, “port knocking and other uses of ‘recent match.’ ” www.shorewall.net/portknocking.html (accessed aug. , ) . ibid. . openwrtdocs/installing­openwrt, wiki.openwrt.org/ openwrtdocs/installing (accessed aug. , ). index blending is the process of database development whereby various components are merged and refined to create a single encompassing source of information. once a research need is determined for a given area of study, existing resources are examined for value and possible contribution to the end product. index blending focuses on the quality of bibliographic records as the primary factor with the addition of full text to enhance the end user’s research experience as an added convenience. key examples of the process of index blending involve the fields of communication and mass media, hospitality and tourism, as well as computers and applied sciences. when academia, vendors, subject experts, lexicographers, and other contributors are brought together through the various factors associated with index blending, relevant discipline-specific research may be greatly enhanced. a s consumers, when we set out to make a purchase, we want the utmost in quality, and when applica­ ble, quantity, and of course all of the other ”appeal” factors that might be associated with a given product or service. these factors may include any number of catego­ ries, not the least of which is price. in other words, let it suffice to say that, as buyers, we want to have our cake and eat it, too. but how often is this a realistic approach to evaluating a given item for purchase? we first must decide what is important to us, decipher the order of this importance as we see it, and evaluate our options. wouldn’t it be much easier if one product in every situ­ ation had all of the factors that we deem important, and the appropriate price to go along with it? according to veliyath and fitzgerald in an article published in competitiveness review, firms can either posi­ tion themselves at the high end, offering higher quality at higher prices, or at the lower end, offering lower quality at a lower price (or anywhere in­between on the continuum of constant value for customers). customers, however, want more of what they value, such as convenience, speed, state­of­the­art design, quality, etc. competitors then try to differentiate themselves from their rivals along the same line of constant value, either by offering a higher quality at the same price or the same quality at a lower price (thereby increasing value for the customer). as such, and using a common example, is it possible to have the handling of a bmw sports car, the luxurious ride of a cadillac, the passenger space of a winnebago, the cargo space of an oversized pick­up truck, all for the price of an economy car? it’s doubtful. but through recent developments in the electronic research database market­ place, and a process known as “index blending,” we may be closer than ever to this ideal formula when it comes to web­based reference resources for academic libraries. the phrase “index blending” is used here to describe an original concept/methodology initiated by ebsco publishing (ebsco). this is not to say that ebsco is the first vendor ever to have combined resources to create a new product, but to the authors’ best knowledge, no other vendor has pursued the “blending” of resources to the same extent and with such a strong guiding directive as ebsco has. index blending is the combining of niche indexes and other important components to create a single defini­ tive index for a particular discipline. as vendors seek to offer the most powerful research database for a given area of study, the pieces may come together through a combination of existing resources and proprietary development. in other words, in order to refine the tools used for research in a discipline, existing resources may be combined, fleshed out, further expanded upon, and enhanced to culminate in the archetypical index for the particular discipline. perhaps this represents the solution to the dilemma that “database choices become increas­ ingly complex when multiple sources exist that cover the same discipline.” the idea may seem elementary, but the process, however, can be arduous. processes involved with index blending expand upon the basic development stages asso­ ciated with creating a research database from “scratch,” coupled with an increase in applicable factors, which become evident when several existing and emerging resources are involved and subsequently interwoven. as is always the case, the first step to building a solution is to identify the problem and/or the need. in database devel­ opment, this is, in a nutshell, pinpointing a subject area of research that is lacking a corresponding definitive index, and where study patterns and research interest dictate a need for such a resource. this involves not only conduct­ ing surveys and engaging in discussion with advisory boards, librarians, subject experts, users, etc., but also taking a close look at the research resources that are cur­ rently available to determine value. because the process begins with the fact that there is a problem (no definitive index for the particular area in question), the idea is to understand the strengths of available resources, as well as to identify weaknesses. through this research process, vendors can further identify independent elements of each resource that may index blending | brooks and herrick index blending: enabling the development of definitive, discipline-specific resources sam brooks and mark herrick sam brooks (sbrooks@ebscohost.com) is the senior vice president of sales & marketing for ebsco information services. mark herrick (mherrick@ebscohost.com) is the vice president of business development for ebsco publishing. � information technology and libraries | june � information technology and libraries | june provide significant benefit or value, as well as pinpoint the additional important pieces that are not represented in any of the available resources. in both cases (available and not available), these elements may represent various aspects associated with a research index such as content coverage (both current and backfile), quality of indexing and abstracts, software/search functionality, thesauri, etc. once the identification and research has taken place, vendors should have the necessary knowledge to proceed to the production phase. figure helps to illustrate how the index blending process can help to develop a new database that fuses together the strengths of existing resources while simul­ taneously compensating for any individual weaknesses that they may have. if value is attributed to currently available databases, then, if appropriate, database acquisition may come into play. this is often a critical phase of the process, and may involve the acquisition of more than a single index. however, the desire by a vendor to acquire a given resource is based on several motivating factors, including the qual­ ity of the database as a whole, the depth and breadth of its coverage, and at times, the extreme quality of an intricate aspect of a database, which will eventually be said data­ base’s contribution to the process of index blending, thus representing its “mark” on the final product. because there is no authoritative resource available for a given subject area does not mean necessarily that certain aspects of existing resources are not of utmost quality. hence, utilizing strengths of existing resources makes sense so as to not “reinvent the wheel” when applicable. in a journal of academic librarianship article discussing the research environment in libraries and the simultaneous utilization of existing library resources, similar principles to those used in index blending are apparent. “properly combining library resources to func­ tion collectively as a cohesive, efficient unit is the basis of information integration.” similar themes to those asso­ ciated with information integration run through index blending. this is attributed largely to the fact that the basic goal of each is to enable the extraction and utiliza­ tion of essential material pertinent to specific research so as to enhance the overall research process. n the process of index blending an example an interesting example of index blending utilized for a major area of study is in the case of communication and mass media. an article in searcher outlined the develop­ ment process and release of the database, communication & mass media complete, which may be the quintessential instance of the power brought about through index blending. in the article, the author first identifies the problem/need as such: when a communication studies student approaches my reference desk, it can take a few moments before i choose a database to search. why the delay? well, to be perfectly blunt, the communication studies literature is all over the place. if the question relates to an aspect of the communications industry, i will often begin with a business database. if the question concerns the effects of media violence on children, i may choose to search one or more of the following: comabstracts, psychinfo [sic], sociological abstracts, eric, and even a few large aggregators, such as wilsonweb’s omnifile and ebsco’s academic search premier. in addition, there is the question of finding a single database that covers the communication science and disorders field and the more mass media­focused communication studies field. the result has been a searching strategy that relies on consulting multiple databases—a strategy that may not please impatient or inexperienced patrons. the need for such an assortment of databases is symptomatic of the discipline. the field of com­ munication studies is extremely interdisciplinary. the discipline’s roots began in the study of rhetoric and journalism and now encompass subjects ranging from political communication to film studies to advertising to journalism to communication disorders to digital convergence and to every manner of media. the dis­ cipline has strong roots in the social sciences, but also draws heavily on the humanities and the sciences. as some have put it, there is an aspect of communication studies in every discipline. this leaves librarians with the difficult task of finding a single database that cov­ ers this wide­ranging discipline. enter ebsco’s new communication & mass media complete database. figure . the index blending process public libraries and internet access | jaeger, bertot, mcclure, and rodriguez �index blending | brooks and herrick � this overview of the need for a comprehensive resource in areas related to communication and mass media is indicative of the type of information that vendors must extract when deciding their course of action for creat­ ing (or not creating) a database to meet such needs. in this instance, the need became apparent to ebsco upon conducting investigative research in this direction. there were certainly important, quality resources available cov­ ering some of the subject areas and subdisciplines, but not a single, all­encompassing resource. hence, the table was set to move forward and begin the process of data­ base development using the process of index blending. once the need for a comprehensive communication and mass media database was established, ebsco began the phases of looking closely at available resources and gathering specific important details about what was required to develop such a database. in order to under­ stand the finer details and make appropriate forward progress in formulating an index for a given research area, a dedicated group of subject experts (advisory board, indexers, lexicographers, etc.) must be estab­ lished. in addition, aggregators must develop appro­ priate relationships and key partnerships. in the case of the database communication & mass media complete, ebsco worked diligently to assemble a panel of experts to provide direction. often, suggestions made by advi­ sory board members ultimately led to larger organiza­ tional partnerships. the first of ebsco’s major partnerships for the benefit of the development of communication & mass media complete was with the national communication association (nca). nca is the oldest and largest national organization to promote communication schol­ arship and education. founded in , the nca is a nonprofit organization of approximately , educa­ tors, practitioners, and students who work and reside in every u.s. state and more than twenty countries. the purpose of the association is to promote study, criti­ cism, research, teaching, and application of the artistic, humanistic, and scientific principles of communica­ tion. nca is a scholarly society and, as such, works to enhance the research, teaching, and service produced by its members on topics of both intellectual and social significance. staff at the nca national office follows trends in national research, teaching, and service pri­ orities. it relays those opportunities to its members and represents the academic discipline of communication in those national efforts. in addition to providing insight and advice into the areas associated with communication and mass media, nca found in ebsco an ideal partner to further the tremendous efforts the organization had put into its database, commsearch. commsearch, in its original form, was a scholarly communication database with deep, archival coverage of the journals of the nca and other major journals in the field of communication studies. the database provided bibliographic and keyword references to twenty­six journals in communication studies with coverage extending to the inaugural issue of each—some from as far back as the early decades of the twentieth century. the database also included cover­to­cover indexing of the nca’s first six journals (from their first editions to the present) and author­sup­ plied abstracts from their earliest appearance in nca journals. as ebsco’s goals were in line with the nca in terms of improving scholarly research in areas sur­ rounding communication as well as enhancing the dis­ semination of applicable materials, a partnership was formed, and ebsco acquired commsearch. the com­ pany acquired this database with the intent to enhance the collection through content additions such that it would take residence immediately as a core component of communication & mass media complete. the second major database acquisition came about similarly to the commsearch arrangement; only this time, ebsco worked closely with penn state university, the developers of a database called mass media articles index. created by jack pontius and maintained by the penn state libraries since , mass media articles index provided citation coverage for over forty thousand articles on mass media published in over sixty research journals, as well as major journalism reviews, recent encyclopedias, and handbooks in the area of communications studies. this database, which was once a stand­alone research tool, is a good example of how a good­quality resource can arise out of the passion and unique vision of an individual, yet never fully develop into its full potential due to a lack of funding, dedicated staff, and experience in database publishing. seeing the incredible potential of mass media articles index, ebsco earmarked this database as the sec­ ond major component in its larger communication and mass media product. as mentioned, the basic idea with index blending is to pinpoint the best and most important aspects of each database to carry forward into the final product. it is at this point that difficulty typically arises in the normalization of data. once core database components are determined, a vendor ’s expertise in building data­ bases, standardizing entries, etc., comes to the forefront. furthermore, because another basic ingredient to the process of index blending revolves around additional material included by the database developer, that aggre­ gator has the burden of taking the core building blocks of the database and elevating these raw materials to the point where their combination and refinement become the desired end result—a definitive, cohesive index to research in the subject area. with this in mind, ebsco carefully selected the indexing components of each resource that were essen­ tial to carry forward and substantially expanded the information technology and libraries | june information technology and libraries | june abstracting and indexing coverage of appropriate journals in commsearch and mass media articles index. the company also added indexing and abstracts for many more of the important titles in the communication and mass media fields that were not covered by these databases. through its initial research, ebsco gained a thorough knowledge of which journals and other content sources were not covered by the two acquired databases, and worked to provide coverage for those missing sources. as such, the idea with this database was to cover all appropriate, qual­ ity titles indexed in all other currently avail­ able communication and mass media­specific databases combined, as well as other important journals not previously covered by any such database. further still, the company took the database to new levels through the creation and deployment of features such as searchable cited references and index browsing. figure provides a visual interpretation of the elements associated with this particular example of index blending. often academic librarians consider aggre­ gated full­text databases as a means for access­ ing full­text information quickly, but with a negative outlook toward the quality of the indexing included in these databases. however, it is ebsco’s intention to create first and fore­ most a powerful index, such that any full text included is that much easier to locate and utilize. according to cleveland and cleveland in the book introduction to indexing and abstracting, rd ed., “in any retrieval system, success or failure depends on the adequacy of the indexing and the related search­ ing procedures.” ebsco wholeheartedly agrees with this statement. and though the company is the leader in providing full­text databases, it continues to raise the bar for these databases through not only constantly increasing the quality and quantity of full text, but also by enhancing indexing, abstracts, and associated search functionality. a database may provide the greatest collection of full text, yet it is still only as good as its underlying indexing framework that guides users to the appropriate content. index blending allows for this ideal because the development of the indexing takes place at the onset as the primary objective, and full text may be included at a later stage. this is precisely the case with ebsco’s communication/communications database where the first iteration of the collection (communication & mass media index) did not include full text, and the complete (full­text version) was soon to follow. thus, in the case of communication & mass media complete, once the core elements for the index were in place, refined, and normalized, ebsco moved forward in the area of full­text content. in addition to the inclu­ sion of full text for all of the nca journals, which david oldenkamp refers to as “heavyweights in communication studies,” ebsco included full­text coverage for nearly titles. according to oldenkamp, as of april , the competing database with the next largest number of publications covered in full text included only sixteen full­text titles. though index blending is not the traditional way in which to build a database, and may actually be the most labor­intensive way in which to proceed, the end results can be remarkable when done properly. using this process, “ebsco has managed to create the largest and most comprehensive database serving the needs of communication studies scholars, faculty, students, and librarians.” in addition, a review published in the charleston advisor determined that “ebsco has brought together two reliable but atrophied resources and refreshed them with new search capabilities and added content, such as abstracts. these have been combined with a healthy dose of ‘not indexed anywhere’ new titles and interdisciplinary sources to create a comprehensive figure . indexing components of communication & mass media complete public libraries and internet access | jaeger, bertot, mcclure, and rodriguez index blending | brooks and herrick resource that will satisfy the needs of students, faculty, and researchers.” n another example of index blending hospitality & tourism index index blending is a concept as much as it is a process and a means to an end. much like applying a particular theory to a number of different instances, index blending is inter­disciplinary in application. thus, the area of com­ munication/communications as described previously, is simply an example of practical implementation of this concept, and a particular way in which the process was approached given the specific elements involved. another discipline to which index blending has been applied is the niche areas related to hospitality and tourism. according to professor vivienne sario, director of travel and tourism at community college of southern nevada, “on a global basis the hospitality and tourism industry employs more than percent of the worldwide workforce. it contributes over $ trillion in gross global output. this means travel and tourism is the world’s largest industry.” though still considered (perhaps incorrectly) a “niche” area of study, the number of hospitality and tour- ism programs supported in colleges and universities around the globe has also increased to the point where dozens and dozens of two- and four-year academic institutions provide related courses of study. from a business perspective, in order to justify the amount of resources that would inevitably be expended to develop a high-end, comprehensive database, the basic criteria needed for database development must first be in place. considering the economic vastness of the hospitality and tourism industry, the interest and research need is quite apparent. if there is at least one clearly definitive academic resource covering the subject area, in all likelihood, the decision would be made to cease exploration and devel- opment in that area. contrarily, when ebsco conducted exhaustive research to determine the need for a new index to literature in the areas of hospitality and tourism, the unani- mous conclusion was to move forward in the development of a product that would go above and beyond the level of the existing resources. this is not to say that quality was not inherent in some of the existing resources. in actual- ity, the fact that there were already quality (albeit perhaps incomplete) resources available, paved the way for utilizing principles of index blending in the development of a more comprehensive resource. the first element of what was to become ebsco’s hospitality & tourism index was purdue university’s lodging, restaurant, & tourism index (lrti). as an indi­ cator of the level of emphasis attributed to this subject area by the university, purdue’s hospitality and tourism management undergraduate program was ranked num­ ber one nationally by a survey published in the journal of hospitality & tourism education. a previous survey conducted by the same journal used a different method­ ology and sample, but still ranked purdue’s hospitality and tourism management (htm) program number one in the nation. to provide insight into the purdue htm program, the origins and history of lrti, the need for a compre­ hensive database, and the university’s decision to work with ebsco, questions were asked of two prominent purdue faculty members: raphael kavanaugh, head, hospitality and tourism management department, and priscilla geahigan, head, consumer and family sciences library. the following is taken from e­mail cor­ respondence among one of the authors (sam brooks), kavanaugh, and geahigan: brooks: how long has purdue offered a hospitality & tourism management program? kavanaugh: the program began in as the department of institutional management. brooks: when and why did purdue decide to create the lodging restaurant & tourism index (lrti)? kavanaugh: to fill a serious void of access to relevant research conducted related to the industry. geahigan: before coverage of the hospitality industry within business indexes and databases was limited. to meet the needs of researchers and students, purdue’s restaurant, hotel, institutional, and tourism management department, an in­house indexing project, started in the purdue consumer and family sciences library in . citations of articles from scholarly and trade journals were entered on index cards, filed by subject headings. in the project became more for­ malized and migrated into partnership with a few other academic institutions. a printed index titled lodging and restaurant index started. in , purdue became the sole producer of the index. in , the index was renamed the lodging, restaurant, and tourism index (lrti), with expanded scope and coverage. over the years, data diskettes and cd­rom formats were added to the printed version. brooks: how important are “niche” or subject­specific databases to support research in a given area such as h&t? geahigan: in contrast to earlier years, students can now get their information from a multitude of databases and information technology and libraries | june information technology and libraries | june venues. at purdue, we have databases that cover all aspects of business and management. undergraduate students often get confused and impatient at the large number of databases offered. a subject specific database like hti gives them a place to start without feeling lost. brooks: why did purdue decide to partner with ebsco, and subsequently merge lrti in the larger hospitality & tourism index (hti)? geahigan: we realized that we do not have the resources to support a database that measures up to industry technology standards and have long decided to look for a company to take over lrti. ebsco’s offer was attrac­ tive to purdue because of their willingness to assume future indexing of the lrti journals. in addition, many purdue students are already familiar with the ebsco interface because we have numerous other ebsco hosted databases. we are pleased that lrti became the foundation of ebsco’s building of hti. the second foundational component of the database also came about through acquisition from an academic institution. articles in hospitality and tourism was copro­ duced by oxford brookes university and the university of surrey. bournemouth university was also a source of data for this database between the years of and . this database provided details of more than forty­six thousand english­language articles selected from more than relevant academic and trade journals published worldwide from to . rounding out the list of three existing resources that were acquired by ebsco, the hospitality database (acquired from the original developers at cornell university) was also assimilated into the new hospitality and tourism database. the hospitality database evolved from the print publication bibliography of hotel management and related subjects that was originally established in the s by blanche fickle, the first director of the library at cornell university’s school of hotel administration. this database, founded on the vision of ms. fickle, would serve as a core resource for ebsco’s new hospitality & tourism index by providing it with a foundation of quality indexing for journals related to the study of hotel adminis­ tration and management. ebsco completed the initial development of its hospitality and tourism database by reviewing applicable subscription statistics maintained by its sister company, ebsco subscription services, in order to locate other publications relevant to the various subdisciplines of hospitality and tour­ ism. any such publications that were not already indexed by the other three existing resources were targeted for inclusion in the new hospitality & tourism index. figure provides a visual interpretation of the ele­ ments associated with this particular example of index blending. following the initial release of hospitality & tourism index, in order to provide an even more inclusive research experience, ebsco proceeded to develop and release a full­text version of this resource entitled hospitality & tourism complete. this new variant of the database offers users the same indexing infrastructure as hospitality & tourism index, as well as provides the additional benefit of immediate access to relevant full­text content. while the availability of full text is certainly of immense value, it is still the quality of underlying indexing that allows this database to be regarded as truly innovative. in fact, this same perspective was echoed in a recent review in choice where the author states that “hospitality & tourism complete indexes its specialized subject area bet­ ter than any other product currently available.” n the whole is greater than the sum of its parts the process of index blending not only brings together content from a variety of resources, it also has the power to increase the research value of that same content. by combining such content under the umbrella of a single comprehensive database, pertinent information can now be more efficiently accessed and cross­referenced with other relevant content. previously, the same body of information could only be explored via a highly ineffec­ tive, piecemeal research process. one last example that demonstrates this potential increase in research value is found in the computers & figure . indexing components of hospitality & tourism complete public libraries and internet access | jaeger, bertot, mcclure, and rodriguez index blending | brooks and herrick applied sciences complete database. this resource was shaped through the acquisition and merger of three distinct indexes—computer science index (csi), internet & personal computing abstracts (ipca), and information science & technology abstracts (ista)—and rounded out with addi­ tional indexed content relevant to the larger discipline. this resulted in a total of , active journals indexed back as far as . then, after two years of dedicated licensing work with pub­ lishers, full text for more than of those titles was added to provide more direct access to such content for researchers. figure illustrates how the various subject areas (unique and shared) covered by the three original databases were merged together in the blending process. from this diagram, it is apparent that the original three databases were already quality resources in their own right and adequately rep­ resented their respective subject areas. however, it should also be apparent that, through the pro­ cess of index blending, the value of the original databases has been enhanced via the fusion of their unique, yet complementary content into a single comprehensive resource. n conclusion though the above examples of communication & mass media complete, hospitality & tourism index, and computers & applied sciences complete represent only three of sev­ eral subject­specific databases culminating from the process of index blending, most database producers (including ebsco) would likely agree that this is not a common procedure for database development. however, the knowledge that a company derives from the pro­ cess often has a significant impact on the company’s other, “nonblended” databases. index blending typically requires a high degree of refinement in order to be fully successful, so when a company engages in this rigorous developmental process, the newfound experience and expertise gained from it may spill over into the com­ pany’s other database initiatives. end users may notice improved indexing, abstracts, and other valuable com­ ponents that are now included in other more established full­text resources from the same vendor. databases that were once viewed simply as “aggregated full­text data­ bases” may be looked upon in a different light after the company adopts the process of index blending for other, unrelated database projects. though these databases may still provide easy access to an abundance of full­text content, they may also now be considered the definitive index for their respective subject area(s). therefore, when a company implements the practice of index blending for some of its products, the resulting effects are two­ fold. the databases created directly as a result of the index blending process are the first to benefit, and the company’s other databases (including those with full text) may also benefit from index blending in an indirect manner. in the end, however, the success of any index blending initiative is measured by the level of benefit that it provides to applicable researchers and other users of the resulting databases. references . rajaram veliyath and elizabeth fitzgerald, “firm capabil­ ities, business strategies, customer preferences, and hypercom­ petitive arenas: the sustainability of competitive advantages with implications for firm competitiveness,” competitiveness review ( ): – . . m. suzanne brown, jana s. edwards, and jeneen lasee­ willemssen, “a new comparison of the current index to jour­ nals in education and the education index: a deep analysis of indexing,” the journal of academic librarianship (may ): – . . sam brooks, “integration of information resources and collection development strategy,” the journal of academic librarianship (july ): – . . david oldenkamp, “ebsco’s new communication and mass media complete (cmmc) database,” searcher , no. (apr. ): . . national communication association web site. http:// www.natcom.org (accessed aug. ). figure . subject areas of component databases are merged into a cohesive whole through index blending information technology and libraries | june information technology and libraries | june . donald b. cleveland and ana d. cleveland, introduction to indexing and abstracting, rd ed. (greenwood village, colo.: libraries unlimited, ): . . oldenkamp, “ebsco’s new communication and mass media complete (cmmc) database.” . ibid. . dodie owens, “advisor reviews—standard review: communication and mass media complete,” the charleston advisor , no. (apr. ): . . vivienne sario, “hospitality & tourism programs,” http://www.studyusa.com/articles/hospitality.asp (accessed june , ). . purdue university web site. http://news.uns.purdue. edu/uns/html ever/ .kavanaugh.rank .html (accessed june , ). . michael g. brizek and mahmood a. khan, “ranking of u.s. hospitality undergraduate programs: – ,” journal of hospitality & tourism education , no. ( ): . . raphael kavanaugh and priscilla geahigan, e­mail mes­ sage with author sam brooks, feb. , . . articles in hospitality and tourism web site (hosted by the university of surrey). http://libweb.surrey.ac.uk/aht /about .asp (accessed june , ). . cornell university’s school of hotel administration web site. http://www.nestlelib.cornell.edu/history.html (accessed june , ). . s. c. awe, “reference­social and behavioral sciences— hospitality & tourism complete,” choice , no. (june ). information technology and libraries | march m any things happen on the national front that affect libraries and their use of technology. legislative action, national policy, and stan­ dards development are all arenas in which ala and lita both take an active role. lita has articulated in its strategic plan the need to pursue active involvement in providing its expertise on national issues and standards development. lita achieves these important objectives in a variety of ways. lita has several committees, interest groups, and representatives to ala standing committees that address legislation, regulation, and national policy issues that pertain to technology. the charge of the lita legislation and regulations committee reads: “the legislation and regulation committee monitors legislative and regula­ tory developments in the areas of information and communications technologies; identifies relevant issues affecting libraries and assists in developing appropri­ ate strategies for responding to these issues.” as its educational mission, the committee publicizes issues and strategies on the lita web site. the chairperson of this committee serves as the lita representative to the ala legislation assembly which advises ala on positions to take regarding legislative and regulatory action. lita also has a representative to the ala office of information technology policy advisory committee who works closely with the legislation and regulation committee on it policy issues that may cross over into the legislative realm. lita also appoints a representa­ tive to the ala intellectual freedom committee whose purpose is “to recommend such steps as may be neces­ sary to safeguard the rights of library users, libraries, and librarians, in accordance with the first amendment to the united states constitution and the library bill of rights.” much has happened on the national front in the past few years that provides plenty of work for these lita and ala committees. the patriot act, calea, net neutrality, dopa, ada compliance, and debates over copyright and intellectual property rights in an electronic world are all examples of issues that require technologi­ cal control or affect systems and network solutions. they also touch at the heart of what librarians have always stood for: protection of intellectual property, personal pri­ vacy, and intellectual freedom. library technologists exert enormous time and effort protecting the privacy of patron records through data retention policies, system controls, and strong authentication systems all while providing authorized access to intellectual property according to copyright or licensing restrictions. keeping lita mem­ bers apprised of all of these issues and the technologies required to abide by legal requirements is an enormous task of the committees and interest groups. these groups do this through programming, publications, and postings to the lita web site. lita has always been very active on the standards development front. from the start, lita was involved with the marc standards through the hard work of henriette avram. the number of standards that affect libraries has mushroomed. there are standards for all aspects of technology—data formats, hardware and firmware, and networking. ala regularly calls on lita to provide expertise on developing standards that per­ tain to library technology. lita has a standards interest group and shares membership with alcts and rusa on the marbi committee. most lita interest groups deal with standards of some sort at least occasionally. the lita board felt that lita’s work on develop­ ing standards was so important that in a new standards coordinator position was created and diane hillman, cornell university, was appointed as the first person in this role. the standards coordinator identifies lita experts to assist in calls for review of developing standards and seeks input from the membership. the standards coordinator works closely with the standards interest group to help educate the membership. because of the nature of digital information, networks, and the standards that enable the distribution of digital informa­ tion and services, it has become impossible for any one person to understand all the standards that affect the library technologist. as standards proliferate, it becomes more important for lita to provide educational oppor­ tunities alongside the involvement in the development of these standards that so impact our daily lives. the lita web site provides a wealth of information about standards. a new means of contributing to the dialogue about developing standards is to participate in the lita wiki where diane hillman will be leading the way in posting information about various library technology standards. also, a great place to learn about various stan­ dards is right here in ital. practically every issue has at least one article about one standard or another. lita’s participation in technological developments on the national front is critical to all libraries. policy, regu­ lation, and standards form the infrastructure to techno­ logical implementation and are the cornerstone to library technology. lita is the place where you can learn more about these developments and participate in the dialogue about them. bonnie postlethwaite (postlethwaiteb@umkc.edu) is lita president / and associate dean of libraries, university of missouri–kansas city. president’s column bonnie postlethwaite is ital’s th volume. my th birthday was the occasion of a great deal of bizarre behavior by my work colleagues, who booby­trapped my office. i do not like cake but love radishes. my birthday “cake” at work was a cheese ball decorated with forty radishes stuck on toothpicks. since i didn’t have to blow them out, i ate them—all forty. ital’s fortieth is no time for such shenanigans. rather it is a time for reflection, celebration, and memoriam. fred kilgour, the founding editor of the journal of library automation (jola), ital’s original title, died last summer. in planning for the th anniversaries of lita in and ital in , the editorial board and i wanted to honor fred as founding editor. i called him and invited him to submit an article of his choosing. he thanked me but graciously declined. he was busy writing his mem­ oirs and said that he needed to conserve his strength for that task. to honor him as founding editor, i have invited a number of authors to submit articles describing their research or their seminal thoughts on our profession. readers have, i hope, seen those articles that are so des­ ignated by notes. i have also invited all lita members to submit such articles in previous editorials and in a posting to lita­l. several articles have resulted from these invitations. this being the first issue of the volume, it is neither too late for me to reissue an invitation, nor too late for you lita members and ital readers to respond with articles that commemorate our fortieth. i’m old enough to know that it is a cliché to proclaim “there has never been a more exciting time to be a librar­ ian.” it was so when volume of jola appeared in . it is so today. let us together peruse the tables of contents (tocs) of the first two issues. vol. , no. ned c. morris, “computer based acquisitions system at texas a&i university”; richard d. johnson, “a book catalog at stanford”; robert wedgeworth, “brown university library fund accounting system”; richard e. chapin and dale h. pretzer, “comparative costs of converting shelf list records to machine readable form”; richard de gennaro, “the development and administration of automated systems in academic libraries” vol. , no. lawrence auld, “automated book order and circulation control procedures at the oakland university library”; donald v. black, “creation of computer input in an expanded character set”; frederick c. kilgour, “costs of library catalog cards produced by computer”; r. a. kennedy, “bell laboratories’ library real­time loan system (bellrel)” four things are immediately striking about those titles. their authors described computer­based solutions and systems for big issues facing libraries forty years ago. second, those problems were all administrative, i.e., they involved using computers to increase the productivity of major operations performed by librarians and library staff. to paraphrase an oft­cited goal, they were systems designed to attempt to control the rate of rise of library costs of operations—to improve the efficiency and effec­ tiveness of internal library processes. therefore third, they were not systems for library users per se. and fourth, they were harbingers of success. global cooperative cataloging and well­integrated library systems have revolutionized our operations. we are devoting relatively more resources to direct services than we did forty years ago. i do not mean that no thoughts or efforts were being devoted to improved user services. when these articles were published, lockheed and the system development corporation (sdc) were in the process of developing the first commercially successful, general online database search systems. in fact, forty years ago, in a former life, as it were, i was present at what i believe was the first trans­ continental online information search, from a teletype machine in sdc’s office in dayton, ohio, to a computer at its santa monica headquarters. (aside to readers: as an impatient young man, i was struck less by the “magic” of the event than by an observation that i expressed on the spot: the response time was horrible—unacceptable. i opined that no one would put up with such a wait. i narrowly escaped with my scalp intact.) the national library of medicine (nlm) was perfecting the medical literature analysis and retrieval system (medlars), medline’s (medlars online’s) predecessor. selective dissemination of information (sdi) services were already being provided using batch processes. computers gen­ erated a myriad of printed article and technical report indexes. we’ve come a long way in forty years. an article in the current issue describes what librarians need to know about “facebook.” increasingly, in information­rich soci­ eties, our students and others want and need their infor­ mation technology on the run. the first five paragraphs of this editorial were com­ posed three weeks ago using the word processor on my palm treo whilst i sat in medical­center waiting and examining rooms in portland, oregon. i downloaded the tocs of jola to my home desktop computer in vancouver, washington, two weeks ago. yesterday, i editorial: reflections on forty john webb john webb (jwebb@wsu.edu) is a librarian emeritus, washington state university, and editor of information technology and libraries. editorial | webb contiuned on page information technology and libraries | march information technology and libraries | march . if you answered “yes” to question , please describe how facebook could be considered an aca­ demic endeavor. ______________________________________________ ______________________________________________ ______________________________________________ ______________________________________________ . please check all answers that best describe what effect, if any, use of facebook in the library has had on library services and operations?  has increased patron traffic  has increased patron use of computers  has created computer access problems for patrons  has created bandwidth problems or slowed down internet access  has generated complaints from other patrons  annoys library faculty and staff  interests library faculty and staff  has generated discussion among library faculty and staff about facebook . is privacy a concern you have about students using facebook in the library?  yes  no  not sure please list any observations, concerns, or opinions you have regarding facebook use in libraries. extracted the paragraphs from my palm to my desktop, and saved that document and the tocs on a universal serial bus (usb) key. today, i combined them in a new document on my laptop and keyed the remaining paragraphs in my room at an inn on a pier jutting into commencement bay in tacoma on southern puget sound. i sought inspiration from the view out my window of the water and the fall color, from old crow medicine show on my ipod, and from early sixties beyond the fringe skits on my treo. fred kilgour was committed to delivering informa­ tion to users when and where they wanted it. libraries must solve that challenge today, and i am confident that we shall. editorial continued from page in march the university of mississippi libraries made our metasearch tool publicly available. after a year of working with this product and integrating it into the library web site, a wide variety of libraries interested in our implementation process and experiences began to call. libraries interested in this product have included consor- tia, public, and academic libraries in the united states, mexico, and europe. this article was written in an effort to share the recommendations and concerns given. much of the advice is general and could be applied to many of the metasearch tools available. google scholar and other open web initiatives that could impact the future of metasearching are also discussed. m any libraries are looking for ways to facilitate the discovery process for users. implementing a one­stop search product that does not require database­specific knowledge is one of the paths librar­ ies are choosing. as these search engines are made available to patrons, the burden of design falls to the library as well as to the product developers. most library users may be familiar with a few databases, but the vast majority of electronic resources remain unrevealed. using a metasearch product, a single search is broadcast out to similar and divergent electronic resources, and search results are returned and typically mixed together. metasearch results are returned in real­time and link the user to the native interface. although there are many products that support one­stop searching, the university of mississippi libraries chose to purchase innovative interfaces’ metafind product because it tied into a digital initiative partnership with innovative. some of the possibilities of the types of resources you can search include: n library catalogs n licensed databases n locally created databases n full text from journals and newspapers n digital collections n selected web sites internet search engines the simplicity of google searching is very appeal­ ing to users. in fact, users have come to expect this kind of empowering tool. at the university of mississippi, students use and have been using google for research. as google scholar went public, it became evident that university faculty also use it for the same reasons. it was apparent from the university of mississippi libraries’ libqual+ survey results that users would like more personal control than the library was offering (table ). unintentionally elaborate mazes are created and users become lost in a quagmire of choices. as indicated by our libqual+ survey results, our users want easy­to­use tools that allow them to find informa­ tion on their own, and they want information to be easily accessible for independent use. these are clearly two areas that many libraries are struggling to improve for their patrons. the question is how to go about it. based on several changes made between and , which included implementing a metasearch tool, the adequacy mean improved for both questions and for undergradu­ ates as well as graduate students and faculty (table ). the adequacy mean compares the minimum level of ser­ vice that a user expects with the level of service that they perceive. in table , the negative adequacy mean figures indicate that the library was not meeting users’ minimum level of service for these two questions or that the per­ ceived level of service was lower than the minimal level of service. table compares the adequacy mean from with and indicates a notable, positive change in adequacy mean for each question and with each group. n design perspectives and tension generally, there are conflicts within libraries regarding the question of how to improve access for patrons and allow for independent discovery. for those leading a metasearch implementation, these tensions are important to understand. in implementing new technologies, there are key development issues that may decrease internal acceptance until they are addressed. however, one may also find that there are some underlying fears regarding this technology. although the following cross­subculture comparisons simply do not do justice to each of the valid perspectives, these brief descriptions highlight the types of perspectives one might encounter when considering or implementing a metasearch product. expert searchers prefer native interfaces and all of the functionalities of the native interface. they are typically unhappy with the “dumbed­down” or clunky searching of a metasearch utility. they would prefer for patrons to be taught the ins and outs of the database they should be using for their research. this presupposes that the students either know which database to use, will spend time inves­ tigating each database on their own, or that they will ask for assistance. however, there are clearly native interface information technology and libraries | june metasearching and beyond: implementation experiences and advice from an academic library gail herrera gail herrera (gherrera@olemiss.edu) is assistant dean for technical services & automation and associate professor at the university of mississippi. metasearching and beyond | herrera functionalities—such as lim­ iting to full text—that, while wonderful to patrons, are not consistent across resources or a part of the metasearch standard. users would cer­ tainly benefit if limiting to full­text was ubiquitous among vendors and if there were some way to determine full­text availability within metasearch tools. results ranking is another issue that expert searchers may bring to the table. currently, there is a niso metasearch initiative that is striving to standard­ ize metasearching. another downside for the expert searcher is that there is no browse function. those who are in administrative or manage­ rial positions working with electronic resources see metasearching as an opportunity to reveal these resources to users who might not otherwise discover them. for example, many users have learned to search ebsco’s academic search premier not realizing that key articles on a local civil rights figure such as james meredith are also available in america: history & life, jstor, and lexisnexis. metasearching removes the need for the user to spend additional time choosing databases that seem relevant and searching them indi­ vidually. from a financial perspective, if a library is pay­ ing for these electronic resources, they should be using them as much as possible. and while the university of mississippi libraries generally target the undergraduate audience with our metasearch tool, the james meredith search is a good example of how a metasearch tool might reveal other databases with information that a serious researcher could then further investigate by link­ ing through the citation to the native interface. those associated with library instruction may also be uncomfortable with metasearching. in fact within a short time of implementing the product, several instructors conveyed their fear that in making searching so simple, they would no longer have a job as the product developed. generally, it seems that users are always in need of instruc­ tion although the type of instruction and the tools continue to change. it is an understandable fear and one that would be wise to acknowledge for those embarking on a metasearch implementation. while metasearch can be an empowering tool for users, you may also encounter some emotional reactions among library employees. from an information literacy point of view, frost has noted that metasearching is “a step backward” and “a way of avoiding the learning process.” it is true that in providing an easy search tool, the library is not endeavoring to teach all students intermedi­ ate or advanced information retrieval knowledge or skills. however, it is important to provide tools that meet users at their level of expertise and as previously noted, this is an area identified in need of improvement. for those working at public service points such as the reference desk, metasearching is an adjustment. many times those working with patrons tend to use databases with which they are more familiar or in which they feel more confident. federated search tools may reveal resources that are typically less used and therefore unfa­ miliar to library employees. training may then become an issue worthy of addressing not just for the metasearch interface and design but also for the less­used resources. for those involved in technical support, this product may range from exciting to exasperating. the amount of time your technical support personnel have to dedicate to your metasearch project should be a major factor when investigating the available products. just like any other technological investment, you are either going to ( ) purchase the technology and outsource manage­ ment or ( ) obtain a lesser price from a vendor for the tool and invest in developing it yourself. there is also a middle ground, but this cost­shifting is important to keep in mind. regardless of your approach, it is critical to include the technical support person on your imple­ mentation team and to keep in mind the kind of time investment that is available when reviewing prices. along with developing this product, one may also find oneself investing additional time and money into infra­ structural upgrades such as the proxy server, network equipment, or dns servers. in addition to these perspectives, there is a general tension in library web site design philosophies between how librarians would like patrons to use their services table . libqual adequacy mean undergrad grad faculty easy-to-use access tools that allow me to find things on my own -. -. -. making information easily accessible for independent use . -. . table . positive change in libqual adequacy mean from to undergrad grad faculty easy-to-use access tools that allow me to find things on my own . . . making information easily accessible for independent use . . . information technology and libraries | june and what patrons want. the traditional design based on educating users and having users navigate to information “our way” has definitely curtailed over the past several years with attention being paid increasingly to usability. as usability studies give librarians increasing informa­ tion, libraries are moving toward designing for our users based on their approaches and needs rather than how librarians would have them work. depending on where one’s library is in this spectrum of design philosophy, implementing a metasearch tool may be harder or easier. judy luther surmised the situa­ tion well, “for many searchers, the quality of the results matter less than the process—they just expect the process to be quick and easy.” moving toward this lofty goal is to some extent dictated by the abilities and inabilities of the technologies chosen. as a technologist, the general rule seems to be that the easier navigation is made for our users; the more complex the technical structure becomes. n metasearch categories in arranging categories of searches for a metasearch product, some libraries group their electronic resources by subject, and others use categories that reflect full­text avail­ ability. the university of mississippi libraries use both. the most commonly used category is our full­text category. this full­text category was set as the default on our most popular search box located on our articles and databases web page (figure ). since limiting to full­text materials is not a standard, the category was defined by the percentage of full­text they contain. this is an important distinction to understand because a user may receive results that are not full­text, but the majority of results will likely be full­text. at our library, if the resource contains more than percent full­text, it is included in the full­text category. other categories included in this implementation are ready reference, library catalogs, digital collections, lim­ ited resources, publicly available databases, and broad subject categories. one electronic resource may be included in the full­text category, a broad sub­ ject category such as “arts and humanities” and also have its own individual category in order to mix and match individual resources on sub­ ject guides using a tailor­made search box. the limited resource category contains resources that should be searchable using the metasearch tool but that have a limited number of simultaneous users. if it were included in the default full­text category that is used so much, it would tie up the resource too much. investigating resources with only one or two simultaneous users at the begin­ ning of the project may help you avoid error messages and user frustration. one might wonder, “why profile limited resources then?” there may be specific search boxes on subject guides where librarians decide to add that individual but limited resource. it might also be necessary to shorten the time­out period for limited user resources. along those same lines, having pay­per­search resources profiled could also be expensive and is not recommended. since the initial implementation, migrating away from per­ search resources has become a priority. within the first few months of implementation, the free resources such as pubmed and askeric were moved to a new “publicly available” category. the reason is that since there is not any authentication involved, these results return very quickly and are always the first results a user sees. while they are important resources, our intent was really to reveal our subscription resources. this approach allows users to search these resources if specifically chosen but they are not included in the default full­text category. this approach does still allow subject librarians to mix and match these free individual resources on subject guide search boxes. n response time of all of the issues with our metasearch tool, response time has been the most challenging. there are so many issues when it comes to tracking down sluggish response that it can be extremely difficult to know where to start. if one’s metasearch software is not locally hosted, response time could involve the library network, campus network, off­campus network provider, and the vendor’s network, not to mention the networks of all the electronic resources users are searching. when one adds the other variable of authentication, the picture becomes even more over­ whelming and difficult to troubleshoot. for authentication, the university of mississippi libraries purchased innovative’s web access management module (wam), which is based on the figure . metasearch tailored search box with full text category selected metasearching and beyond | herrera ezproxy software. as the use of our electronic resources from on­campus and off­campus has grown, the inci­ dence of increasing network issues has risen. in work­ ing with our campus telecommunications group, the pursuit of ever­greater bandwidth has become a priority. troubleshooting has included tracking down trouble­ some switch settings, firewall settings, as well as campus dns and vendor dns issues. if your network adminis­ trators use packet shapers, this may be another hurdle. clearly, our metasearch product has placed a significant load increase on the proxy server. in looking at proxy statistics, percent of total proxy hits were from the metasearch product (figure ). with this in mind, one may find the load on one’s proxy server increasing very dramatically during peak usage and may need to plan for upgrades accordingly. even with improvements and tweaks along the way, response time is still an issue and one of the highest hurdles in selling a metasearch product internally and externally. one metasearch statistical module includes response time information for individual resources along with usage data. the response time information would be very helpful in troubleshooting and in working with electronic resource vendors. usage tracking is another criterion to consider in reviewing metasearch products. n response time and tailored search boxes during implementation, one of the first discussions to have is who will be the target audience for this product. at this institution, undergraduates were the target audi­ ence and more specifically, those looking for three to five articles for a paper. while our metasearch software has a master screen showing all of the resources divided into the main categories, facing users with over sixty check boxes was not a good solution (figure ). this master screen is good for demonstrating categories to library staff, overall functionality of the technology, and also for quickly checking all of your resources for connectivity errors. from early conversations with students, keeping basic users far away from this busy screen is a good goal. remember, the purpose is to give them an easy starting point. the best way to keep users in a simple search box is to construct search boxes and hand­pick either individual resources or categories keep­ ing in mind the context of the web page. for example, the articles and databases page has a simple search box that searches for articles. subject guide boxes search individual electronic resources selected by the subject librarian. the university of mississippi libraries also have a large col­ lection from the american institute of certified public accountants (aicpa). the search box on that page searches our catalog, which contains aicpa books along with the aicpa digital collection. some libraries are interested in developing a standard metasearch box to display as a widget or standing content area throughout their web site. this is interesting and worth considering. however, matching the web page content with appropri­ ate resources has been our approach. as the standards and technology develop, this may be worth further con­ sideration depending on usability findings. for the most commonly used search box on the articles and databases page (figure ), the default category checked is the full­ text articles category. donna fyer stated that, “for the average end user, the less decision making, the better.” this certainly rings true for our users. originally, a simple metasearch search box was placed on the library homepage. the library catalog and the basic metasearch box were both displayed. this seemed confusing for users since both products have search capabilities. with the next web site redesign, the basic metasearch box moved from the library homepage to the articles and journals web page. this was a success­ ful place for the article quick search box to reside since the default was set to search the full­text category. there were some concerns that users might be typing journal titles into the search box but these were rare instances and not necessarily inappropriate uses. the next rede­ sign eventually moved this search box to the articles and databases page, where it remains. for the articles and databases pages, the simple search box (figure ) by default searches the full­text category and searches the title keyword index. the index category with the label, “article citations,” can also be checked by the user. the majority of metasearches begin with this search box and figure . total proxy hits vs. metafind proxy hits � information technology and libraries | june most users do not change the default settings for the resources or the index. n subject guide search boxes in addition to the “article quick search” box, subject librarians slowly became interested in a search box for their subject guides as the possibili­ ties were demonstrated. in order to do this, the ven­ dor was asked to profile each resource with its own unique value in order to mix and match individual resources. while the idea of searching resources by subject category sounds useful and appealing, sometimes universal design begets universal dis­ cord. even with a steering committee involved, it is hard for everyone to agree what resources should be in each of the main subject categories: arts and humanities, science and engineering, business and economics, and social science. some libraries have put a lot of time and effort into creating a large number of subject categories. the master search screen (figure ) displays several of this library’s categories but not the broad subject categories noted above. these general sub­ ject categories are brought out in the multipurpose interface called the “library search engine” (figure ). the library search engine design is a collection of the categories and resources showing the full functionality of our metasearch tool. the subject categorization approach within our metasearch interface is a good way to show the multifunction­ ality of the product but remains relatively unused by patrons. by giving each resource its own value, subject librarians have the flexibility to select spe­ cific resources and/or categories for their subject guides. it is worth noting that it required additional setup from our vendor and was not part of the original implementation. after a few months of testing with the initial implemen­ tation, willing subject librarians chose individual resources for their tailored search boxes. once a simple search box has been constructed, it can be easily copied with minor modi­ fications to make search boxes for those requesting them. while progress was slow to add these boxes to subject guides, after about a year there was growing interest. in setting these up, subject librarians have several choices to make. first of all, they choose the resources that will be searched. for example, the biology subject guide search box searches academic search premier, bioone, and jstor by default. basicbiosis and pubmed are also avail­ able but are not checked by default. users can check these search boxes if they also wish to search these resources. choosing the resources to include in the search box as well as setting what resources are checked by default is the most important decision. the subject librarian is also encour­ aged to assist in evaluating the number of hits per resource returned. with response time being a critical factor, deter­ mining the number of hits per resource should involve testing and take into consideration the overall number of resources being searched. n relevance selecting the default index is another decision in setting up search boxes. again, users are google­oriented and tend to go with whatever is set as the default option. out of the box, our metasearch tool defaults to the keyword index or keyword search. the issue of relevancy is a hot topic for metasearch products. this issue typically comes up in metasearch discussions. it is also listed as an issue in the niso metasearch initiative. from the technical side of the equation, results are displayed to the user as soon as they are retrieved. this allows users to begin immediately exam­ figure . master screen display (partial screenshot) figure . library search engine subject categories metasearching and beyond | herrera � ining the results. adding a relevancy algorithm as a step would mean all of the results would have to be returned, ranked, and then displayed. with response time being a key issue, a faster response is more important than relevance. another consideration is if the metasearch results are displayed to the user as interfiled or by electronic resource where the resource is returning results based on its own relevancy rankings. one way to increase relevance is to change the default index from keyword to title keyword. for our students, bringing back keywords in the title made the results more relevant. this is the default index used for our article search on the articles and database web page. subject librarians have the choice of indexes they prefer when blending resources. one caveat in using title keyword is that there are resources that do not support title keyword searching. for other resources, title keyword is not an appropriate index. for example, wilson biographies does not have a title keyword search. it makes perfect sense that a biography database would not support title keyword searching. in these cases, the search may fail and note that the index is not supported. to accommodate this type of exception, the profile for wilson biographies needed to be changed to have the title keyword search­mapped to a basic keyword search. while this does not make the results as relevant as the other search results, it keeps any errors from appearing and allows results to be retrieved. n results per source and per page for metafind, there are also two minor controls that can work as hidden values unseen by the patron or as compo­ nents within the search box for users to manipulate. the first control is the number of hits to return per resource. if a subject librarian is only searching two or three resources in his tailored search box, he probably will want to set this number higher. if there are many resources, this number should be lower in order to keep response time reasonable. the second control is the number of results to return per page. in general, it is important to adjust these controls after testing the response for the resources selected. while users typically use the default settings, showing these two con­ trols gives the user a visual clue that the metasearch tool is not retrieving all of the results from the resource. instead, it is only retrieving the first twenty­five, for example. n implementation advice one of the most important pieces of advice is that it is extremely important to have a date in one’s contract or rfp for all of the profiling to be completed if the vendor is doing the resource profiling. from this library’s experi­ ence, the profiling of a resource can take a very long time, and this is a critical point to include in the contract. one might also consider adding cost and turn­around time for new resources after the initial implementation to the contract. the more resources profiled, the more useful the product. however, one also needs to pay attention to response time. if the plan is to profile one’s own resources or connectors, librarians should be mindful of the time involved and ask other libraries with the same product about time investments. being able to work with vendors who will provide an opportunity to evaluate the product “live” is preferable. in deciding who to target for an implementation team, consider representatives from reference, collection development, and systems. it is also very important to include whoever manages electronic resource access/ subscriptions and a web manager. in watching other pre­ sentations, exclusion of any of these representatives can seriously undermine the implementation. buy­in is essen­ tial to success. additionally, giving librarians as many options as possible, such as control over what types of resources are in their search boxes as well as the number of hits per resource makes the product more appealing. n questions to ask once the implementation team is set, interviewing refer­ ences for the products under consideration is an impor­ tant part of the process. unstructured conversations with references really allow librarians to explore together what the group wants and how its needs fit with the services the vendor offers. a survey of questions via e­mail is another possibility. in choosing this method, be sure to leave some room for open comments. regardless of the approach, it is important to spend some time asking ques­ tions. provided are a list of recommended questions: n who is responsible for setting up each resource—the vendor or you? n how much time does it typically take to set up a new resource and what is the standard cost to add a new resource? n is there a list or database of already­established pro­ files for electronic resources for this product? n how much time would you estimate that it took to implement the product? n will you be able to edit all of the public web pages yourself or will you be using vendor support staff to make changes? if the vendor support staff has to make some of the changes, how responsive are they? information technology and libraries | june n can you easily mix and match individual resources for subject guides, departmental pages, or other kinds of web pages? or do you only have the option to set up global categories? n is your installation local or does the vendor host it? are there response issues? n is there an administrative module to allow you to maintain categories, resource values, and configura­ tion options? n how much time goes into managing the product monthly? and who manages the product at your library? n what kind of statistical information does the vendor provide? n how satisfied are you with the training, implementa­ tion support, and technical documentation? n how does the vendor handle broken resources or subscription changes? as with most technologies, there are upfront and hid­ den costs. it is important to determine what hidden costs are involved and if you have the resources to support all of the costs. sometimes libraries choose the least expen­ sive product. however, this approach can lead librar­ ies down the path of hidden costs. for example, if the product is less expensive but your library is responsible for setting up new electronic resources, managing all of the pages, and finding ways to monitor and troubleshoot performance outside of the tools provided, the hidden expenditures in time and training may be more costly in the end than purchasing the premium metasearch tool. in essence, one must pay for the product one way or another. the big question is, where are the resources to support the product? if one’s library has more it/web personnel than money, the lower­costing product may be the way to go, but be sure to check with other librar­ ies to see if they have been able to successfully clear this hurdle. additionally, if your library has more one­time money than yearly subscription money, this may dictate the details of the rfp, and your library may lean toward a purchase rather than an annual subscription. n metasearch summary clearly, students want a simple starting place for their research. implementing a metasearch tool to meet this need can be a hard sell internally for many reasons. at this institution, response time has been the overriding critical issue. response has lagged due to server and network issues that have been difficult to track down and improve. however, authentication is truly the most time­ consuming and complex part of the equation. some fed­ erated search tools are actually searching locally stored information, which helps with response. while these are not truly metasearch tools and are not performing real­ time searches, this approach may yield more stability with faster response. over the years in implementing new services such as the library web site, illiad, electronic resources, and off­ campus authentication, new services are often adopted at a much faster rate by library users than by library employees. typically, there will be early adopters who use the services immediately based on need. it then takes general users about a year to adopt a new service. iii’s metasearch technology has been available for the past four years. however, our implementation is evolving with each web site redesign. still, it is used regularly. the university of mississippi libraries has been pro­ viding access to its electronic resources in two distinct ways: ( ) providing urls on web pages to the native interface of the electronic resource and ( ) metasearching. as the library moves forward in developing digital col­ lections and the number of electronic resources profiled for metasearching increases, it is possible that this kind of global discovery tool will compete in popularity with the library catalog. providing such information mining tools to patrons will cause endless frustration for the library literate. response times, record retrieval order, as well as licensing and profiling issues, are all obstacles to pro­ viding a successful metasearch infrastructure. retrieval inconsistency and ad hoc retrieval order of records is very unsettling for librarians. however, this is the kind of tool to which web users have become accustomed and certainly seems to fill a need that to date has been lacking where library electronic resources are concerned. n open web developments one other trend appearing is scholarly research discovery tools on the open web. enter google scholar along with other similar initiatives such as windows live academic search. google scholar beta was released in november and very soon after began an initiative to work with libraries and their openurl resolvers. this bridging between an open web tool and libraries is an interest­ ing development. a fair amount has been written about google scholar to date although the project is still in its beta phase. what does google scholar have to do with metasearching? good question. it remains to be seen how much scholarly information will become search­ able via google scholar. for now, the jury is still out as to whether google scholar will begin to encroach upon the traditional territory of the indexing and abstracting world. if sufficient content becomes available on the open web, whether from publishers or vendors allowing their metasearching and beyond | herrera content to be included, then the authentication piece that directly effects response time may be overcome. in using google scholar or other such open web portals, search­ ing happens instantly. when a user uses the openurl resolver to get to the full­text, that is where authentication enters into the picture and removes the negative impact on searching. the tradeoff is that there are many issues involved in openurl linking and the standardization of the metadata needed to provide consistent access. there are many parallels between what google scholar is attempting to offer and what the promises of metasearching have been. for metasearching, under­ graduate students looking for their three to five articles for a paper are considered our target audience. for in­ depth searching, metasearching does have limitations, but for the casual searcher looking for a few full­text articles, it works well. interestingly, similar recommen­ dations are being made for google scholar. however, opinions differ on this point. roy tennant went so far as to indicate it is a step forward in access to those users without access to licensed databases, but remained reserved in his opinion regarding the usefulness for those with access. google scholar also throws in a few bonuses. while providing access to open access (oa) materials in our opac for specific collections such as the directory of open access journals, these same resources have not been included in our metasearch discovery tool. google scholar is searching these open repositories of scholarly informa­ tion, although there is some concern over the automatic inclusion of materials such as syllabi and undergraduate term papers within the institutional repositories. google scholar also provides a useful citation feature and rel­ evancy. google scholar recognizes the user’s preference for full­text access and provides a visual cue from the brief results when article full­text is available. this func­ tionality is not currently available from our metasearch software but would be extremely helpful to users. on the downside, some of google scholar’s linking policies make it difficult for libraries to extend services beyond full­ text articles to their users. another notable development among subscription indexing services is the ability to reveal content to web search engines. ebsco’s initiative is called ebscohost connection. in implementing metasearching, libraries have debated about providing access to free versus subscrip­ tion resources. for our purposes, free resources were not included in the most commonly used search in the full­ text category. there are those who would argue against this decision, and they have very good points. in fact, it has already been noted that some libraries use google scholar to verify incomplete interlibrary loan citations quickly. in watching the development of google scholar, it seems possible that this free tool that uncovers free open access resources and institutional repository mate­ rials may not necessarily be a competitive product, but may be a very complementary one. n impact on the opac what will this mean for the “beloved” opac? for a very long time, users have expected more of the library catalog than it has provided. while the library catalog is typically appreciated by library personnel, its usefulness for finding materials other than books has been hard for general users to understand. many libraries including the university of mississippi have been loading records from their electronic resources in hopes of making the library catalog more useful. the current conversation regarding digital library creation also begs the question, “what is the library catalog?” although the library catalog serves as a searchable inventory of what the library owns, it is simply a pointing mechanism, whether it points the user to a shelf, a building, or a url. in our endeavor to provide instant gratification and full­text, as well as the user’s desire for information regardless of format, the library catalog is beginning to take a backseat. it was clear four years ago in plan­ ning digital collections that a metasearch tool would be needed to tie together subscription resources, digital collections, publicly available resources, and the library catalog. it will be interesting to see whether patrons choose to use the formal tools provided by the library or the informal tools developing on the open web, such as google scholar, to perform their research. more than likely, discovery and access will happen through many avenues. while this may complicate the big picture for those in library instruction, it is important to meet users on the open web. one’s best intentions and designs are presented to users but they may choose unintended paths. librarians should watch the paths they are taking and build upon them. sometimes even one’s best attempts fall short, as pointed out clearly in karen schneider’s latest series, “how opacs suck.” still it is important to acknowl­ edge design shortcomings and keep forging ahead. dale flecker, who spoke at the taiga forum, recommended not to spend years trying to “get it right” before imple­ menting, but instead to consider ourselves in perpetual beta and simply implement and iterate. in other words, do not try to make the service perfect before implement­ ing it. most libraries do not have the time and resources to do this. instead, find ways to gain continual feedback and constantly adjust and develop. students are familiar with internet search engines and do not want to choose between resources. access to a simple resource discovery tool is an important service for users. unfortunately, authentication, product design information technology and libraries | june and management, and licensing restrictions tend to be stumbling blocks to providing fast and comprehen­ sive access. regarding the metasearch tool used at the university of mississippi libraries, development part­ nerships have already been formed between the vendor and a few libraries to improve upon many of the issues discussed. innovative is developing a next­generation metasearch product called research pro that leverages ajax technology. while efforts are made to participate in discussions and develop our already­existing tools, it is also impor­ tant to pay attention to other developments such as google scholar. at this point, google scholar is in beta but this kind of free searching could turn the current infra­ structure on its ear to the benefit of patrons. the efforts to meet users on the open web and reveal scholarly content are definitely worth keeping an eye on. references . roland dietz and kate noerr, “one­stop searching bridges the digital divide,” information today , no. ( ): s . . niso metasearch initiative, http://www.niso.org/ committees/ms_initiative.html (accessed may , ). . william j. frost, “do we want or need metasearching?” library journal , no. ( ): . . judy luther, “trumping google? metasearching’s prom­ ise.” library journal , no. ( ): . . donna fyer, “federated search engines,” online , no. ( ): . . jill e. grogg and christine l. ferguson, “openurl link­ ing with google scholar,” searcher , no. ( ): – . . mick o’leary, “google scholar: what’s in it for you?” information today , no. ( ): – . . roy tennant, “is metasearching dead?” library journal , no. ( ): . . o’leary, “google scholar.” . what is ebscohost connection?, http://support.epnet .com/knowledge_base/detail.php?id= (accessed may , ). . laura bowering mullen and karen a. hartman, “google scholar and the library web site: the early response by arl libraries,” college & research libraries , no. ( ): – . . karen g. schneider, “how opacs suck,” ala tech- source, http://www.techsource.ala.org/blog/karen+g./sch­ neider/ / (accessed may , ). . dale flecker, “my goodness, life is different,” pre­ sentation to the taiga forum, mar. – , , http://www .taigaforum.org/pres/fleckerlifeisdifferenttaiga .ppt (accessed may , ). lita cover , cover , cover index to advertisers information technology and libraries | march this study examines how social scientists arrive at and utilize information in the course of their research. results are drawn about the use of information resources and channels to address information inquiry, the strategies for information seeking, and the difficulties encoun- tered in information seeking for academic research in today’s information environment. these findings refine the understanding of the dynamic relationship between information systems and services and their users within social-scientific research practice and provide implica- tions for scholarly information-system development. t he information needs and information­seeking behavior of social scientists have been the focus of inquiry within library and information science (lis) research for decades. folster reviewed the major studies that have been conducted in this area over the past three decades. she found that research methods had developed through several stages. research prior to the s usually consisted of questionnaire­based user studies that gathered basic demographic data and quan­ titative data on the type of information used. following that were citation studies in the mid­ s, and then the combination of questionnaire and interview techniques to develop profiles of users and their needs in the s. the information environment of the s witnessed a major transition in research design. the former practice of studying large groups via questionnaires or struc­ tured interviews gave way to the use of unstructured interviews or observation of smaller groups, resulting in a more holistic picture of social scientists’ research practices. more fully developed techniques for behavioral models emerged in the s. folster summarized these studies done over decades and concluded that ( ) social scientists place a high importance on journals; ( ) most of their citation identification comes from journals; ( ) infor­ mal channels, such as consulting colleagues and attend­ ing conferences, are an important source of information; ( ) library resources, such as catalogs, indexes, and librar­ ians, are not very heavily utilized; and ( ) computerized services are ranked very low in their importance to the research process. there are many examples of studies about the infor­ mation­seeking behavior of social scientists. for example, the infross project (investigation into information requirements of the social scientist) studied the informa­ tion needs of british social scientists in the late s and early s and found that they preferred to use journal citations instead of traditional bibliographic tools, and that they tended to consult with colleagues and subject experts, rather than library catalogs or librarians in order to locate information. other social­scientist studies reinforced the findings of the infross project. several studies indicated that computerized literature searching was ranked low as a source of information among social scientists and suggested the promotion of electronic information services by librarians to enhance their roles as information providers. in an influential study on social scientists’ informa­ tion­seeking patterns, ellis developed a behavioral model with six features based on the stages they went through in gathering information: ■ starting—includes activities characteristic of the ini­ tial search for information, such as asking colleagues or consulting literature reviews, online catalogs, and indexes and abstracts; ■ chaining—following chains of citations and other forms of referential connection between materials; ■ browsing—semi­directed searching in an area of potential interest, such as scanning published jour­ nals, tables of contents, references, and abstracts; ■ differentiating—using differences (authors or jour­ nal hierarchies) between sources as a filter on the nature and quality of the material examined; ■ monitoring—maintaining awareness of develop­ ments in an area through the monitoring of particular sources such as core journals, newspapers, confer­ ences, magazines, books, and catalogs; and ■ extracting—systematically working through a par­ ticular source to locate material of interest, for exam­ ple, sets of journals, collections of indexes, abstracts, or bibliographies. meho and tibbo revised ellis’s information­seeking model by studying the information­seeking behavior of social­science faculty who study stateless nations. they confirmed ellis’s model and derived four additional fea­ tures—accessing, networking, verifying, and information managing. accessing is getting hold of the materials or sources of information once they have been identified and located. networking includes communicating and maintaining a close relationship with a broad range of people such as friends, colleagues, and intellectu­ als. verifying is checking the accuracy of the informa­ tion found, and information managing includes filing, archiving, and organizing the collected information to facilitate research. yi shen yi shen (yishen@wisc.edu) is a ph.d. candidate in the school of library and information studies, university of wisconsin- madison. her article is the winner of the lita/endeavor student writing award. information seeking in academic research: a study of the sociology faculty at the university of wisconsin-madison article title | author information seeking in academic research | shen with the exception of ellis’s work in – and the follow­up study by meho and tibbo, studies inves­ tigating academic social scientists have been in steady decline since the mid­ s. according to line, in an information world radically changed by the internet, it is essential to carry out new studies of information uses and needs. most of the studies discussed in this paper were conducted before the development of the internet. the present study focuses on the information­seeking behav­ ior of social scientists in a new information environment featuring the internet and other dramatic technological advances. kling and mckim pointed out the growing importance of information technology and the resulting major shifts in scientific practice. costa and meadows studied the impact of computer usage on scholarly com­ munication among social scientists and found that major changes in their communication habits were occurring. the most significant impacts of information technology were greater interactivity, widened community boundar­ ies, extended access to information, and an increasing democratization of the international research community. they suggested that the developments were influenced by new pressures (social, economic, political) from the research community and the institutional environment, and by newly available resources (infrastructure, ser­ vices, sources) being introduced into the academic envi­ ronment by information technology. it could be expected that social scientists’ information­seeking behavior would change within a new social­technical environment. the purpose of this study is to extend the findings of the pre­ vious studies by examining social scientists’ information needs and their activities and perceptions in relation to today’s information systems and services. this paper provides a theoretical framework for the study, discusses the methods for data collection and data analysis, and summarizes findings. finally, it discusses results, reflects on the theoretical and practical implica­ tions that ensue, and notes the limitations imposed by the study design. ■ theoretical framework the theoretical frame for this study is the idea of “com­ munities of practice.” wenger, mcdermott, and snyder define a community of practice as “a group of people who share a common concern, a set of problems, or a passion about a topic, and who deepen their knowledge and expertise in this area by interacting on an ongoing basis.” within communities of practice, people share common values, observe and interact with each other, exchange views and ideas, and contribute to the knowl­ edge­creation process. according to wenger, communities of practice are combinations of three elements: a domain of knowledge, which defines the key issues in the community; a com­ munity of people who care about the domain; and the shared practice that they create. communities of prac­ tice are loosely connected, informal, and self­managed. they are about knowledge sharing, and the best way to share knowledge is through social interaction and infor­ mal relationship networks. effective communication and mutual understanding are important factors in fostering communities of practice. this form of social construction is highly situated and highly improvised. it essentially suggests that researching some thing is inseparable from its own historical and social locations of practice and should be carried out in the process of actually doing that thing. a process organizes knowledge in a way that is especially useful to practitioners whose shared learning brings value to a community. pragmatically, the exami­ nation of context­based research processes draws “atten­ tion away from abstract knowledge and cranial processes and situates it in the practice and communities in which knowledge takes on significance.” what is learned is highly dependent in the context on which the learning takes place, as it is central to the transfer and consump­ tion of information. this requires “looking at the actual practice of work, which consists of a myriad of fine­ grained improvisations that are unnoticed in any formal mapping of work tasks.” such beliefs are utilized in this present study to approach and explain information­seek­ ing behavior among social scientists. researchers used communities of practice in orga­ nization and business studies to investigate knowledge sharing and knowledge­creation processes within orga­ nizational settings to cultivate the building of knowl­ edge­management systems. researchers also used this approach in the field of computer­supported cooperative work (cscw) to study the social interactions of group­ ware systems and community computing and support systems. this study selected communities of practice as the theoretical frame because it has been widely applied in the study of knowledge sharing and has been tested and verified through empirical research. this study rep­ resents an exploration of the usefulness of communities of practice for research on information­seeking behavior within a knowledge­intensive scholarly community. the primary purpose of the present study is to pro­ vide empirical evidence on social scientists’ information seeking in scientific research. the main research ques­ tions are: ( ) how do social scientists make use of different information sources and channels to satisfy their infor­ mation needs? ( ) what strategies do they apply when seeking information for academic research? and, ( ) what difficulties are encountered in searching for supporting information technology and libraries | march information technology and libraries | march information? information service providers should find the results of this study interesting because identifying users’ perceptions of the information environment pro­ vides guidance for information­system development that will closely reflect or accommodate the information­seek­ ing activities of social scientists. ■ methods the research questions described in the preceding section were tested in the context of information use in scientific inquiry by faculty in the department of sociology at the university of wisconsin­madison during march and april . the participants were selected from the faculty list on the department web page and then contacted by e­mail to arrange face­to­face interviews. four people were interviewed based on their willingness to par­ ticipate. three of them are full­time professors and have teaching experience of more than twenty years (one of them has been teaching for more than thirty years). the fourth is an assistant professor with four years of teaching experience. all of the participants are female. each inter­ view lasted from forty­five minutes to an hour. all participants were interviewed in their campus offices to allow for easy access to supporting materials as examples of how they go about their work. after explain­ ing her identity, the purpose of the research, and assuring the confidentiality of the interview, the researcher asked initial questions in a relatively structured way to glean background­related information and research context. the second part of the interview dealt with informa­ tion­related behavior, such as information sources and channels used to address research inquiry, and the major strategies for selecting needed information. the third part focused on problems the participants encountered in information seeking. the researcher took field notes and tape­recorded all interviews. as a consistency check, the participants were sometimes asked to comment on disciplinary work prac­ tices gleaned from other interviews. the selection of four participants reflected the practicalities of collecting data with limited time and resources. ■ findings based on the idea of communities of practice that what is learned is highly dependent on the context in which the learning takes place because it is central to the trans­ fer and consumption of information, the present study provides a holistic picture of information use situated in actual research practice and academic context among these social scientists. these findings can be summarized into several interrelated stages as shown in figure . the figure shows that the social scientists’ information seeking moves from academic information needs, choice of information sources, searching for information, to use of the information. the researchers move back and forth between stages until the information inquiry is satisfied. searching for information involves the implementation of strategies, confrontation of difficulties, and continuous decision making. choice of information channels goes through the whole information­seeking process based on researchers’ momentary or changing information activi­ ties and information needs. this figure is intended to provide a general view of the information seeking behav­ ior in this specific case, but is not intended to generate a model or pattern of information seeking. the findings are organized into the use of information resources and channels to address information inquiry, the strategies for information seeking, and the difficulties encountered in searching for information, which together constitute the major information­seeking practice of the participants. figure . stages of the social scientists’ information seeking article title | author information seeking in academic research | shen ■ use of information resources and channels to address information needs information needs the respondents reported their research­oriented infor­ mation needs in the context of their research activities. those information needs can be grouped into seven cat­ egories. examples of responses follow. . general academic issues and current research dis­ courses in the field. “i find conferences are more useful for seeing what kinds of general things are going on. i guess some of these are research, some are academic politics kinds of things, and what’s happening in the disci­ pline as a whole.” “in conferences, you find out what other people are doing research on. the most current research is not published yet, so you know what’s happening now.” . feedback from colleagues on personal research. “the best thing about conferences is that when i present my own research, i get comments about it.” “you show your paper to people and ask them for comments, and they show you their papers and ask you for comments. this is kind of the normal part of academic life.” “i usually send a copy of a paper or something and get actual comments through e­mail.” . current research topics and activities of specific authors. “i’ll look for key people, and see what they’ve done. . . .” “knowing who is doing what where. . . .” “you sort of inevitably talk about your research with other people doing comparable research and find out what they are doing to keep current to what the different research projects are.” . existing datasets (existing survey research data­ bases) and statistics for secondary data analysis. “there are online statistical sources that i get to put in the papers.” “i use the internet to download all the . . . data that we analyze. . . .” “i do a lot of data research, so i use government sites on the internet, like the science’s bureau, or the national center for health statistics. we also have a little center for demography and ecology library. i use our in­house databases too.” “in social science, there are many existing sets of data. we have something called data and program library service here. they have all kinds of data­ bases that will tell you where there are data sources that have certain variables in them. . . . so you can go and do your own statistical analysis on those data.” . information needed for management purposes, such as the cooperation and coordination of research activities. “in this department, we conduct community busi­ ness by e­mail. we pass messages around. . . . a decision is usually made through this dialogue.” “i am constantly in interaction with people by e­ mail to cooperate on research projects.” . community recognition and inspirational support from colleagues. for example, one respondent commented, “in conferences, i feel invigorated when sitting and talking to field colleagues who are interested in my research. the whole conversa­ tion makes me feel excited and inspired.” another respondent indicated, “to see people face­to­face that you respect and they think your work is good, that’s good.” it is echoed by a third respondent: “you just talk about your work, and people act like what you are doing is very interesting, then it makes you more inspired.” those needs for information constitute a major research practice of the participants and thus determine how they go about seeking information. ■ information resources supporting information resources could be divided into internally built university resources and external resources. moreover, these internal and external resources could be further subdivided into human resources and nonhuman resources based on their physical forms. internal, nonhuman resources the participants identified two major categories of internal nonhuman information resources for academic research based on their intended use. the first of these categories is books and journals that are available in the university libraries for literature reviews and to provide awareness of current research. however, because of phys­ ical inconvenience, campus libraries are not often used. one participant indicated, “the library is down the hill, so even before there were lots of good internet resources, i wasn’t going down to the library a lot.” on the other hand, the participants reported that they frequently used the library online public access catalog (opac) to order information technology and libraries | march information technology and libraries | march document delivery from the libraries. “i find madcat (the library online catalog system) very useful for a whole variety of specific searches for journals, books, and differ­ ent online information.” another participant remarked, “i can request a book online through the document deliv­ ery services.” another internal nonhuman resource consists of exist­ ing survey datasets that are collected by the center for demography and ecology library for secondary data analysis and research. it was indicated that in social sci­ ence, as more and more survey research databases were available, there was an increasing amount of research conducted on secondary data. the data and program library service provides all kinds of databases informing researchers of the location of data sources and the vari­ ables contained in certain datasets. external, nonhuman resources the participants identified three types of external non­ human resources based on their medium. some of these resources are purchased and managed internally by the campus libraries but developed and maintained exter­ nally by outside library and information professionals. one type is electronic resources, such as electronic news­ papers, external opacs, electronic full­text databases, online statistical reports, survey databases, and govern­ ment or personal web sites. some named examples include sociological abstracts, lexis­nexis, science bureau’s web site, the national center for health statistics web site, web of science, and online british newspapers. the second type is printed resources, such as books, newspapers and magazines, archives, and newspaper indexes that are available from outside of the campus. named examples include the paper indexes for the new york times and los angeles times back in the s. the third type is audio­ video resources, such as radio broadcasts, tapes, video­ tapes, and television. one major finding was that the participants depended primarily on electronic information resources. all looked for information on both literature and research data via the internet. literally, each participant had her own fre­ quent visit to search engines or opacs for information on specific research topics and general research subjects. examples of responses include: “i start with internet explorer and go to google.” “i work a lot online. . . . i just do internet searches. “both these journal and newspaper databases, i use a lot for various purposes.” “i want to find out if there is work on this specific topic or concept. i would almost always start with sociological abstracts.” “the citation index is terrific for finding contempo­ rary work building on something important.” moreover, the respondents also conducted research on the internet to study web behavior or social networks on the internet. “there are more and more people actually doing research on the internet, studying web sites or connec­ tions between web sites. . . . they collect data online. . . .” “in social­movement research, more and more researchers study how people coordinate transactional movements, protest movements, various ethnic move­ ments, and political movements through the internet.” “online is a big way of doing cooperation as well as doing research. it is one of the reasons that we are inter­ ested in studying what kind of connections there are on the electronic network.” “a current research project that i am doing is looking at network of . . . web sites. so we are gathering primary data from the web sites.” thus, the electronic mechanism for information sys­ tems and services dominates the manner in which the participants carry out their research. internal, human resources the faculty participants were not only electronic­informa­ tion consumers, but also electronic­information producers. for example, one described, “i maintain my own web page, on which i post my research and add links to outside resources that i collected for years. i have my own gateway to organize the link pages, which can be used for my future reference and by my students. the library links to my web page as well.” moreover, this participant advocated the creation and collection of electronic materials by her col­ leagues as well. “it’s an evolving process. the more people put their information on the internet, the more useful it is to be on the internet. we are right in that transition.” the department can easily take a step further to build a shared pool of information and information resources in its internal system. a second type of internal human resources comes from the technical staff who provided announcements of technical developments and product information, as well as technical assistance for social­science research. working as the social science computing cooperative (sscc), the technical staff provides the faculty with detailed instruc­ tions and useful tips for creating electronic materials as well as with directions for publishing them. librarians, as a third type of human resource, provided reference services and collected necessary information resources for their academic research. external, human resources the external human resources that the respondents gathered and contacted are of two types: people shar­ article title | author �information seeking in academic research | shen � ing similar research interests and concerns, and people having different fields of interest. the former types were valued for supporting suggestive and creative commu­ nication and interaction as well as potential cooperation. for example, “when it comes to really think[ing] about things, sit down in one place and talk, and then stuff comes out. you don’t even know what you are thinking until you sit down and talk to people. it’s idea generating.” “knowing who is doing what where in the field is important. . . . i am working on a . . . research topic, which requires the awareness of other people with similar inter­ est around the world. . . . i cooperate with the scholars from different countries and with different knowledge background.” the latter types are used for current awareness of research works in other fields and general disciplinary activities and academic trends. for example, “i need to know people who know what’s going on in other fields, and they tell me what’s going on.” “i get a lot in terms of contemporary research at con­ ferences, which are useful for things that haven’t been published in journals.” “[a conference] will generate a lot of interesting inter­ change.” “[at conferences], i think about how what other people are doing is related to what i would want to do, or how they can do it differently. a lot of times, i think about whether the methods they are using would be useful for my work at all.” ■ channels the major information channels through which the par­ ticipants delivered and exchanged information included e­mail, telephone, face­to­face communication, and proj­ ect reports or other documents. e­mail was a domi­ nant communication and information­acquisition tool in research. face­to­face or oral communication channels in this case were often used as a supplementary means. “mostly, e­mail is how i communicate with people, occasionally telephone, but not very often. even with people here and we can walk right next door, mostly we just e­mail each other. it’s nice, because you have a record.” “i get hundreds of e­mails a week. . . . i live on e­mail. my colleagues know i am easier to reach by e­mail than in person.” “[face­to face] it’s just the more personal and emo­ tional mode [of communication] . . . you can see the person’s expression, and figure out what they are really thinking.” e­mail communication helped accomplish several scientists’ tasks, including quick exchange of timely infor­ mation, teamwork coordination, non­work­oriented mes­ sage exchange, field discussion, field information seeking and finding communities of interest. for example, one participant indicated the coordination of community activities through e­mail. in this department, we conduct community business by e­mail. community members rarely meet face to face. the chairperson finds out what the research task is, and sends out messages. people exchange opinions through e­mail messages. and a decision is usually made through this dialogue, instead of talking face to face. when scholars are going to have a face­to­face meet­ ing, they deliver the data, records, and reports before­ hand, and share their initial viewpoints with supporting information through e­mail. the following factors affected a scholar’s choice of channels for information delivery and exchange: the char­ acteristics of the information receiver, the characteristics of the information, the task or purpose of delivering or sharing information, and the immediacy of response. for example, one respondent mentioned that she usu­ ally delivered data, records, and research documents via e­mail for formal announcement and record keeping by the receivers. when there was no stress of immediate response, she preferred e­mail communication for the thoughtful input and feedback allowed by the asynchro­ nous­exchange feature of e­mail. “intellectual questions are more easily handled by e­mail because i have the time to think about it and formulate my responses.” she con­ tinued, “i usually e­mail a copy of my paper to colleagues for detailed feedback.” in another case, a participant indicated, “some of us are well aware that e­mail is archived, it’s not anonymous and not private. if you are concerned about something and want to say something that you don’t want to have an e­mail record of, you may want to go to talk to some­ body about it, instead of writing it in an e­mail.” another participant explained that because of her research topics, she usually adopted the face­to­face method of communication and attended all kinds of international academic conferences. in other circum­ stances, when collecting opinions for resolving certain questions, she chose to use e­mail. ■ strategies for information seeking the participants indicated certain strategies applied to gathering information and tracking resources to address information technology and libraries | march information technology and libraries | march their information needs. those strategies with response examples are: . extracting abstracts: “i use abstracts to get the parameters of what’s happening and then know more narrowly where to focus.” . tracking citations: “the citation index is terrific for finding contemporary work” that builds on previ­ ous major work in a subject area. . restricting the search to a limited set of sources or types of sources to achieve satisfactory results within an acceptable timeline. . constantly filtering and interpreting the search results by referring to the summary description of web sites: “in most searches that i do, the first ten hits are book dealers. i don’t bother with them. i go to the next page and try. . . . i look at the summary of what the site is and try to figure out what the worthy things are.” . avoiding search terms prone to commercial infor­ mation: “when searching for something without a lot of commercial stuff, you are more likely to get what you want on the top.” . setting the default for the number of search results with consideration of information completeness, information usefulness, importance of research, and timeliness: for example, one participant stated, “i usually set my least default to a hundred cita­ tions. five hundred is too many, but it depends on what you’re looking for, how much you care about your findings, how much faith you have for the existence of useful information. if you think it’s not worth a minute of your time, you just forget it. but if you are sure it’s there, you just have to keep looking for and work[ing] harder at it.” as shown in the findings, the participants employed certain criteria for evaluation of the information they gathered. those judgment criteria were: importance of research, usefulness, accuracy, completeness, and timeli­ ness. the results imply that to accomplish the research tasks on hand in a fast­paced and distributed digital­ information environment, the practicalities of time and human effort have come into play in the ways in which the participants sought information. ■ difficulties in seeking supporting information the problems encountered by the participants when col­ lecting information through various resources were iden­ tified and are grouped into categories, including: ■ information is scattered in different places and with different qualities; it is difficult to have a complete and valuable picture of a research phenomenon. the participants described this difficulty as “how tricky computerized search is.” ■ there is too much information on the internet to filter, and the current search techniques and ranking tools are not intelligent enough to capture the most relevant information of interest. the participants described trying alternative search strategies as a “game­playing” and “brainstorming” process. ■ no sources of information or mechanisms assist in the identification of people with similar research interests and their activities in the broad virtual space. for example, one participant described: i am trying to find what’s in public debate on con­ troversial topics. and it’s very common to have trouble finding both sides of the debate. i started with diffuse searches on the internet trying to see if i can find the potential academic community and tag into their debate. i basically searched on [the search term] on the whole internet because i had no idea where it would be, who got involved, and how it was formed. when doing [the research] issue, it’s easy to find the people in favor of [a topic], but difficult to find anybody who was an opponent. eventually, i got hundreds of hits [search results], and i had to wade through a lot of proponents to find the opponents. sometimes, it’s an issue to find [an] ethnic minority perspective of a topic. ■ technology upgrades and system integration arouse another concern. as one participant expressed it, “technology is changing [so] fast that lots of com­ puter files from the s are no longer readable. the danger of an information system lies in the tradeoff between the accessibility provided by digitization and the long­term survival of intellectual proper­ ties.” ■ there are no digital sources of information for some historical documents and no retrievable data­ bases for book chapters. one participant noted, “the online strategy is very good for really current stuff, but not for older stuff. the people who started the . . . research were actually writing before the online revolution, so they are not turning up so much in keyword searches online.” another participant also mentioned the inconvenience of using hard­copy indexes for newspapers from the s and archival data that go back to the s and s. ■ discussion this study shows how the ‘communities of practice’ perspective situates the process of using information in the actual practice of scientific research. it provides an information context in which knowledge takes on sig­ article title | author information seeking in academic research | shen nificance. the results provide empirical evidence of the participants’ activities as well as insights into the ways they seek information. in his discussion of user­oriented evaluation and qualitative analysis of information use, ellis emphasized a small­scale qualitative analysis of users’ perceptions of system performance to construct insights into the complex reality of the information environment. he argued that a detailed understanding of the complexity and interaction of information systems and services and their users can be used to explain problems and provide guidance on the development of information systems. the present study is in accord with ellis’s idea by focus­ ing on a specific sample of academic social scientists working in a university setting. the choice of university of wisconsin­madison is based on the grounds of conve­ nience and ease of access. the restriction to one specific sample also avoids the added complexity and compound problems of information use situated in different practice and contexts. ellis also considered the feasibility of interviews to “provide enough information for a detailed and accurate account of the perceptions of the social scientists of their information­seeking activities to be made, and to enable an authentic picture to be constructed of those activi­ ties.” he thought the information­seeking activities of social scientists were too diffuse to carry out triangulation of methods. by applying the interview method, this cur­ rent study complies with ellis’s suggestion. on the other hand, ellis’s information­seeking behav­ ior­model of social scientists presented six generic fea­ tures. these conclusions are far too general for specific application. from the perspective of communities of practice, the current study examines the way social sci­ entists use information in their research practices and specific circumstances; it also presents specific informa­ tion­related behavior, strategies, and difficulties. this study also extends the understanding of the way infor­ mation is used by social scientists in a new information environment with dramatic technical advances. the findings of this study support the conclusions of kling and mckim and costa and meadows by showing the growing importance of information technology and the resulting major shifts in information­seeking practice among social scientists. unlike research findings prior to the s, the social scientists in this study make exten­ sive use of a variety of information sources and channels, primarily electronic­information systems and services, in seeking information. in the new information environ­ ment, these new information mechanisms also presented limitations and difficulties. moreover, many lis researchers have examined users’ relevance criteria in information seeking. great emphasis is given to the “situational dynamism of user­ centered relevance estimation.” situated in their research practices, the present study also identified the social sci­ entists’ applications of certain criteria for evaluation of information. although the small­scale study has limitations for research generalization, the rich description of social sci­ entists’ perspective on the information environment has some practical implications for information­system­and­ service design for academic social scientists. ■ plan for system-to-system integration this study identified technology upgrades and sys­ tem­integration problems existing in current academic information systems. technology was developed and applied without the capability of intergenerational com­ munications and transactions at the cost of intellectual properties. kling and star addressed the same issue that “computerized systems appear like the layers of an archaeological dig, with newer systems built upon older systems with various workplace surveillance capa­ bilities.” they stated that such “legacy systems” are fragile and inflexible for information use and knowledge management. therefore, planning for system integration should be underway. ■ enhance the web resource-retrieval system the study identified the difficulties encountered by fac­ ulty in locating relevant, complete, and valuable informa­ tion effectively and efficiently on the large and dynamic web. an advanced web resource system thus is required that allows web content to be indexed and retrieved more intelligently. moreover, the findings of information­seek­ ing strategies in this case study suggest a one­way user­ system interaction process. there is no interactive query refinement between the user and the system. thus, the users have to brainstorm and play with alternative search strategies in the hope of significant results. to enhance system effectiveness, a relevance­feedback mechanism that takes into account the users’ relevance judgment is thus needed. this mechanism should have a two­way user­system interaction component. ■ construct an internal information system the findings of the study point to a need for a shared pool of information resources in the university of wisconsin– madison department of sociology. through the leverage and reuse of existing internal knowledge assets in the information technology and libraries | march information technology and libraries | march department, this system could help collectively create or gather information resources for cross­reference by colleagues. ■ construct a collaborative information mechanism for the social-scientific community according to the findings, there are no sources of infor­ mation or mechanisms that assist the identification of people with similar research interests and their activi­ ties on the broad virtual space. however, awareness of shared interests and experiences constitutes an important external human resource that is valued for suggestive and creative interaction and for potential cooperation. thus, a collaborative information mechanism for identification with personal academic interests will be helpful. ■ limitations certain limitations inherent in the study need to be acknowledged. due to the time and resource constraints, the study sample includes only four scholars. given this small sample, results cannot be generalized. although ellis mentioned the feasibility of interviews in a user­ oriented study of information use, dependence on a single method has the disadvantages of the restriction of views. for example, interviewer characteristics, expecta­ tions, and verbal idiosyncrasies, and participants’ socially desirable responses are recognized in many studies as potential sources of method biases (podsakoff et al.). if time and resources permit, triangulation of methods—for example, combining interviews with observations and diaries—would increase the level of specificity and justify the validity and reliability of the research results. ■ conclusion drawing upon the idea of communities of practice that what is learned: ( ) is dependant on the context in which the learning takes place, and ( ) is central to the transfer and consumption of information, this study examined the information­seeking behavior of four social scien­ tists. results were drawn about their use of information resources and information channels to meet their infor­ mation inquiries, their strategies for information seeking, and the difficulties encountered in searching for relevant information, situated in the course of their actual scien­ tific research. this work has two primary contributions. first, it provides a rich description of social scientists’ per­ spectives on their research­oriented information­seeking behavior in the context of today’s information environ­ ment. second, it situates information seeking behavior in a socially constructed practice and presents specific features of information seeking. these results will help refine the understanding of the dynamic relationship between information systems and services and their users within scientific research. several areas remain for future research. researchers could make a comparative study of academics in differ­ ent institutional settings. future research could also study the dynamic interaction of information systems and ser­ vices and their users within each stage of ellis’s model of information­seeking patterns among social scientists to get insights into the specific features of their information seeking behaviors and to enrich their general patterns of information inquiry with specific details. research on information­seeking behaviors of social scientists could also focus on specific research tasks or certain research stages to decide differences or similarities of informa­ tion­seeking behaviors across academic practice. similar research could also be done on faculty in other disci­ plines. references . m. b. folster, “information­seeking patterns: social scien­ tists,” the reference librarian , no. / ( ): – . . m. b. line, “information requirements in the social sci­ ences: some preliminary considerations,” journal of librarianship , ( ): – ; m. b. line, “the information uses and needs of social scientists: an overview of infross,” aslib proceedings , ( ): – . . p. stenstrom and r. b. mcbride, “serial use by social sci­ ence faculty: a survey,” college and research libraries ( ): – ; r. h. epp and j. s. segal, “the acls survey and aca­ demic library service,” college and research libraries news , ( ): – ; m. slater, “social scientists’ information needs in the s,” journal of documentation , no. ( ): – ; m. b. folster, “a study of the use of information sources by social science researchers,” the journal of academic librarianship , no. ( ): – ; c. c. gould and m. j. handler, information needs in the social sciences: an assessment (mountain view, calif.: research libraries group, ). . folster, “a study of the use of information sources by social science researchers”; epp and segal, “the acls survey and academic library service.” . d. ellis, “the derivation of a behavioral model for infor­ mation retrieval system design” (ph.d. diss., univ. of sheffield, ); d. ellis, “a behavioral approach to information retrieval system design,” journal of documentation , no. ( ): – . . l. i. meho and h. r. tibbo, “modeling the information­ seeking behavior of social scientists: ellis’s study revisited,” article title | author information seeking in academic research | shen journal of the american society for information science and technol- ogy , no. ( ): – . . h. c. hobohm, “social science information and docu­ mentation: time for a state of the art?” inspel , no. ( ): – . . m. b. line, “social science information: the poor rela­ tion,” ifla journal , no. ( ): – . . r. kling and g. mckim, “not just a matter of time: field differences and the shaping of electronic media in supporting scientific communication,” journal of the american society for information science , no. ( ): – . . s. costa and j. meadows, “the impact of computer usage on scholarly communication among social scientists,” journal of information science , no. ( ): – . . e. wenger, r. mcdermott, and w. m. snyder, cultivating communities of practice: a guide to managing knowledge (boston: harvard business sch. pr., ), . . s. al­hawamdeh, knowledge management: cultivating knowledge professionals (oxford: chandos pubs., ). . e. wenger, communities of practice: learning, meaning, and identity (cambridge: cambridge univ. pr., ). . j. s. brown and p. duguid, “organizational learning and communities of practice: toward a unified view of working, learning, and innovation,” organization science , no. ( ): – . . j. s. brown, “internet technology in support of the con­ cept of ‘communities of practice’: the case of xerox,” account- ing, management, and information technologies , no. ( ): – ; brown and duguid, “organizational learning and com­ munities of practice; f. blackler, “knowledge, knowledge work, and organizations: an overview and interpretation,” organi- zation studies , no. ( ): – ; j. lave and e. wenger, situated learning: legitimate peripheral participation (cambridge: cambridge univ. pr., ); n. hayes and g. walsham, “par­ ticipation in groupware­mediated communities of practice: a socio­political analysis of knowledge working,” information and organization , no. ( ): – . . wenger, mcdermott, and snyder, “cultivating communi­ ties of practice.” . brown and duguid, “organizational learning and com­ munities of practice,” . . hayes and walsham, “participation in groupware­medi­ ated communities of practice,” . . k. grosser, “human networks in organizational informa­ tion processing,” in m. e. williams, ed., annual review of informa- tion science and technology (medford, n.j.: learned information, ), – ; brown, “internet technology in support of the concept of ‘communities of practice’”; brown and duguid, “organizational learning and communities of practice”; black­ ler, “knowledge, knowledge work, and organizations”; lave and wenger, situated learning: legitimate peripheral participation; hayes and walsham, “participation in groupware­mediated communities of practice.” . d. ellis, “user­oriented evaluation and qualitative anal­ ysis of patterns of information use,” in d. bawden, user-ori- ented evaluation of information systems and services (brookfield, vt.: gower, ), – . . ibid., . . kling and mckim, “not just a matter of time”; costa and meadows, “the impact of computer usage on scholarly com­ munication among social scientists.” . c. l. barry, “user­defined relevance criteria: an explor­ atory study,” journal of the american society for information science , no. ( ): – ; h. w. bruce, “a cognitive view of the situational dynamism of user­centered relevance estimation,” journal of the american society for information science , no. ( ): – ; s. mizzaro, “relevance: the whole story,” jour- nal of the american society for information science , no. ( ): – ; x.­j. yuan, n. j. belkin, and j.­y. kim, “the relation­ ship between ask and relevance criteria,” in proceedings of the th annual international acm sigir conference on research and development in information retrieval (new york: acm pr., ), – ; s. y. rieh, “judgment of information quality and cognitive authority in the web,” journal of the american society for information science and technology , no. ( ): – ; c. n. wathen and j. burkell, “believe it or not: factors influencing credibility on the web,” journal of the american society for infor- mation science and technology , no. ( ): – ; a. tombros, i. ruthven, and j. m. jose, “searchers’ criteria for assessing web pages,” in proceedings of the th annual international acm sigir conference on research and development in information retrieval (toronto: acm pr., ), – . . bruce, “a cognitive view of the situational dynamism of user­centered relevance estimation,” . . r. kling and l. star, “human­centered systems in the perspective of organizational and social informatics,” comput- ers and society , no. ( ): – . . p. m. podsakoff et al., “common method biases in behav­ ioral research: a critical review of the literature and recom­ mended remedies,” journal of applied psychology , no. ( ): – . information technology and libraries | march article title: subtitle in same font author name and second author author id box for column layout information technology and libraries | march article title: subtitle in same font author name and second author author id box for column layout based on data collected as part of the public libraries and the internet study, the authors assess the degree to which public libraries provide sufficient and quality bandwidth to support the library’s networked services and resources. the topic is complex due to the arbitrary assignment of a number of kilobytes per sec- ond (kbps) used to define bandwidth. such arbitrary definitions to describe bandwidth sufficiency and quality are not useful. public libraries are indeed connected to the internet and do provide public-access services and resources. it is, however, time to move beyond connectiv- ity type and speed questions and consider issues of band- width sufficiency, quality, and the range of networked services that should be available to the public from public libraries. a secondary, but important issue is the extent to which libraries, particularly in rural areas, have access to broadband telecommunications services. t he biennial public libraries and the internet studies, conducted since , describe public library involve­ ment with and use of the internet. over the years, the studies showed the growth of public­access comput­ ing (pac) and internet access provided by public libraries to the communities they serve. internet connectivity rose from . percent to essentially percent in less than ten years; the average number of public access computers per library increased from an average of two to nearly eleven; and bandwidth rose to the point where percent of public libraries have connection speeds of greater than kbps (kilobytes per second) in . this dramatic growth, replete with related information technology challenges, occurred in an environment of challenges—among them budgetary and staffing—that public libraries face in main­ taining traditional services as well as networked services. one challenge is the question of bandwidth suf­ ficiency and quality. the question is complex because typically an arbitrary number describes the number of kbps used to define “broadband.” as will be seen in this paper, such arbitrary definitions to describe band­ width sufficiency are generally not useful. the federal communications commission (fcc), for example, uses the term “high speed” for connections of kbps in at least one direction. there are three problematic issues with this definition: . it specifies unidirectional bandwidth, meaning that a kbps download, but a much slower upload (e.g., kbps) would fit this definition; . regardless of direction, bandwidth of kbps is neither high speed nor does it allow for a range of internet­based applications and services. this inad­ equacy will increase significantly as internet­based applications continue to demand more bandwidth to operate properly. . the definition is in the context of broadband to the single user or household, and does not take into consideration the demands of a high­use multiple­ workstation public­access context. in addition to connectivity speed, there are many ques­ tions related to public library pac and internet access that can affect bandwidth sufficiency—from budget and sus­ tainability, staffing and support, to services public librar­ ies offer through their technology infrastructure, and the impacts of connectivity and pac on the communities that libraries serve. one key question, however, is what is qual- ity pac and internet bandwidth for public libraries? and, in attempting to answer that question, what are measures and benchmarks of quality internet access? this paper provides data from the public libraries and the internet study to foster discussion and debate around determining quality pac and internet access. bandwidth and connectivity data at the library outlet or branch level are presented in this article. the band­ width measures are not systemwide but rather at the point of service delivery in the branch. ■ the bandwidth issue there are a number of factors that affect the sufficiency and quality of bandwidth in a pac and internet service context. examples of factors that influence actual speed include: ■ number of workstations (public­access and staff) that simultaneously access the internet; ■ provision of wireless access that shares the same con­ nection; ■ ultimate connectivity path—that is, a direct connec­ tion to the internet that is truly direct, or one that goes through regional or other local hops (that may have aggregated traffic from other libraries or orga­ nizations) out to the internet; john carlo bertot and charles r. mcclure assessing sufficiency and quality of bandwidth for public libraries john carlo bertot (jbertot@fsu.edu) is the associate director of the information use management and policy institute and professor at the college of information, florida state university; and charles r. mcclure (cmcclure@ci.fsu.edu) is the director of the information use management and policy institute (www .ii.fsu.edu) and francis eppes professor of information studies at the college of information, florida state university. article title | author assessing sufficiency and quality of bandwidth for public libraries | bertot and mcclure ■ type of connection and bandwidth that the telecom­ munications company is able to supply the library; ■ operations (surfing, e­mail, downloading large files, streaming content) being performed by users of the internet connection; ■ switching technologies; ■ latency effects that affect packet loss, jitter, and other forms of noise throughout a network; ■ local settings and parameters, known or unknown, that impede transmission or bog down the delivery of internet­based content; ■ range of networked services (databases, videoconfer­ encing, interactive/real­time services) to which the library is linked; ■ if networked, the speed of the network on which the public­access workstations reside; and ■ general application resource needs, protocol priority, and other general factors. thus, it is difficult to precisely answer “how much bandwidth is enough” within an evolving and dynamic context of public access, use, and infrastructure. putting public­access internet use into a more typi­ cal application­and­use scenario, however, may provide some indication of adequate bandwidth. for example: ■ a typical three­minute digital song is mb; ■ a typical digital photo is about mb; and ■ a typical powerpoint presentation is about mb. if one person in a public library were to e­mail a powerpoint presentation at the same time that another person downloaded multiple songs, and another was exchanging multiple pictures, even a library with a t line ( . mbps—megabytes per second) would experience a temporary network slowdown during these operations. this does not take into account many other new high­ bandwidth­consuming applications such as cnn stream­ ing­video channel; uploading and accessing content to a wiki, blog, or youtube.com; or streaming content such as cbs’s webcasting the ncaa basketball tournament. an increasingly used technology in various settings is two­way internet­based video conferencing. with an installed t line, a library could support two kbps or three kbps videoconferences, depending on the amount of simultaneous traffic on the network—which, in a public access context, would be heavy. indeed, the public libraries and the internet study indicated a near continuous use of public­access workstations by patrons (only . percent of public libraries indicated that they always had a sufficient number of workstations available for patron use). public libraries increasingly serve as access points to e­government services and resources, e.g., social services, disaster relief, health care. these services can require the simple completion of a web­based form (low­bandwidth consumption) to more interactive services (high­band­ width consumption). and, as access points to continuing education and online degree programs, public libraries need to offer adequate broadband to enable users to access services and resources that increasingly can depend on streaming technologies that consume greater bandwidth. ■ bandwidth and pac in public libraries today as table demonstrates, public libraries continue to increase their bandwidth, with . percent of public libraries reporting connection speeds of kbps or greater. this compares to . percent of public libraries reporting connection speeds of greater than kbps in . there are disparities between rural and urban pub­ lic libraries, with rural libraries reporting substantially fewer instances of connection speeds of greater than . mbps in . on the one hand, the increase in con­ nectivity speeds between and is a positive step. on the other, . percent of public libraries report that their connection speeds are insufficient to meet patron demands all of the time, and . percent indicate that their connection speeds are insufficient to meet patron demands some of the time. thus, nearly half of public libraries indicate that their connection speeds are insuf­ ficient to meet patron demands some or all of the time. in terms of public access computers, the average number of workstations that public libraries provide is . (table ). urban libraries have an average of . workstations, as compared to rural libraries, which report an average of . workstations. a closer look at bandwidth and pac for the next sections, the data offer two key views for analysis purposes: ( ) workstations—divided into libraries with ten or fewer public­access workstations and libraries with more than ten public­access worksta­ tions (given that the average number of public­access workstations in libraries is roughly ten); and ( ) band­ width—divided into libraries with kbps or less and libraries with greater than kbps (an arbitrary indicator of broadband for a public library context). in looking across bandwidth and public­access work­ stations (table ), overall . percent of public libraries have connection speeds of less than kbps while . percent have connection speeds of greater than kbps. a majority of public libraries— . percent—have ten or fewer workstations, while . percent have more than ten workstations. in general, rural libraries have fewer workstations and lower bandwidth as compared to sub­ urban and urban libraries. indeed, . percent of urban information technology and libraries | march information technology and libraries | march libraries with fewer than ten workstations have connec­ tion speeds of greater than kbps, as compared to . percent of rural libraries. when examining pac capacity, it is clear that public libraries have capacity issues at least some of the time in a typical day (tables through ). only . percent of public libraries report that they have sufficient numbers of workstations to meet patron demands at all times (table ), while nearly as many, . percent, report that they consistently are unable to meet patron demands for public­access workstations (table ). a full . percent indicate that they are unable to meet patron demands during certain times in a typical day (see table ). in other words, . percent of public libraries report that they are unable to meet patron demand for public­access workstations some or all of the time during a typical day—regardless of number of workstations available and type of library. the disparities between rural and urban libraries are notable. in general, urban libraries report more difficulty in meeting patron demands for public­access workstations. of urban public libraries, . percent report that they consistently have difficulty in meeting patron demand for workstations, as compared to . percent of suburban and . percent of rural public libraries (table ). by contrast, . percent of urban libraries report sufficient workstations to meet patron demand all the time as compared to . percent of rural libraries (table ). when reviewing the adequacy of speed of connectiv­ ity data by the number of workstations, bandwidth, and metropolitan status, a more robust and descriptive pic­ table . public library outlet maximum speed of public-access internet services by metropolitan status and poverty metropolitan status poverty level maximum speed urban suburban rural low medium high overall less than kbps . % ± . % (n= ) . % ± . % (n= ) . % ± . % (n= ) . % ± . % (n= ) . % ± . % (n= ) . % ± . % (n= ) . % ± . % (n= ) kbps– kbps . % ± . % (n= ) . % ± . % (n= ) . % ± . % (n= , ) . % ± . % (n= , ) . % ± . % (n= ) . % ± . % (n= ) . % ± . % (n= , ) kbps– kbps . % ± . % (n= ) . % ± . % (n= ) . % ± . % (n= ) . % ± . % (n= , ) . % ± . % (n= ) - . % ± . % (n= , ) kbps– kbps . % ± . % (n= ) . % ± . % (n= ) . % ± . % (n= , ) . % ± . % (n= , ) . % ± . % (n= ) - . % ± . % (n= , ) kbps– . mbps . % ± . % (n= ) . % ± . % (n= , ) . % ± . % (n= , ) . % ± . % (n= , ) . % ± . % (n= ) . % ± . % (n= ) . % ± . % (n= , ) greater than . mbps . % ± . % (n= , ) . % ± . % (n= , ) . % ± . % (n= , ) . % ± . % (n= , ) . % ± . % (n= ) . % ± . % (n= ) . % ± . % (n= , ) don’t know . % ± . % (n= ) . % ± . % (n= ) . % ± . % (n= ) . % ± . % (n= ) . % ± . % (n= ) . % ± . % (n= ) . % ± . % (n= ) weighted missing values, n= , table . average number of public library outlet graphical public- access internet terminals by metropolitan status and poverty* poverty level metropolitan status low medium high overall urban . . . . suburban . . . . rural . . . . overall . . . . * note that most library branches defined as “high poverty” are in general part of library systems with multiple branches and not single building systems. by and large, library systems connect and provide pac and internet services systemwide. article title | author assessing sufficiency and quality of bandwidth for public libraries | bertot and mcclure ture emerges. while overall, . percent of public librar­ ies indicate that their connection speeds are adequate to meet demand, some parsing of this figure reveals more variation (tables through ): ■ libraries with connection speeds of kpbs or less are more likely to report that their connection speeds are insufficient to meet patron demand at all times, with . percent of rural libraries, . percent of suburban libraries, and . percent of urban libraries so reporting (table ). ■ libraries with connection speeds of kpbs or less are more likely to report that their connection speeds are insufficient to meet patron demand at some times, with . percent of rural libraries, . per­ cent of suburban libraries, and . percent of urban libraries so reporting (table ). ■ libraries with connection speeds of greater than kbps also report bandwidth­sufficiency issues, with . percent of rural libraries, . percent of suburban libraries so reporting; and . percent of urban librar­ ies indicating that their connection speeds are insuf­ ficient all of the time (table ); . percent of rural libraries, . percent of suburban libraries, and . percent of urban libraries indicating that their connec­ tion speeds are insufficient some of the time (table ). ■ libraries that have ten or fewer workstations tend to rate their bandwidth as more sufficient at either kbps or less or greater than kbps (tables , , and ). thus, in looking at the data, it is clear that libraries with fewer workstations indicate that their connection speeds are more sufficient to meet patron demand. table . public library public-access workstations and speed of connectivity by metropolitan status rural suburban urban lt kbps gt kbps lt kbps gt kbps lt kbps gt kbps or fewer workstations . % n= , . % n= , . % n= . % n= , . % n= . % n= more than workstations . % n= . % n= , . % n= . % n= , . % n= . % n= , total . % n= , . % n= , . % n= , . % n= , . % n= . % n= , missing: . % (n= , ) table . fewer public library public-access workstations than patrons wishing to use them by metropolitan status rural suburban urban total or fewer workstations . % n= . % n= . % n= . % n= , more than workstations . % n= . % n= . % n= . % n= total . % n= . % n= . % n= . % n= , missing: . % (n= ) table . fewer public library public-access workstations than patrons wishing to use them at certain times during a typical day by metropolitan status rural suburban urban total or fewer workstations . % n= , . % n= , . % n= . % n= , more than workstations . % n= , . % n= , . % n= . % n= , total . % n= , . % n= , . % n= , . % n= , missing: . % (n= ) table . sufficient public library public-access workstations avail- able for patrons wishing to use them by metropolitan status rural suburban urban total or fewer workstations . % n= , . % n= . % n= . % n= , more than workstations . % n= . % n= . % n= . % n= total . % n= , . % n= . % n= . % n= , missing: . % (n= ) information technology and libraries | march information technology and libraries | march table . public library connection speed insufficient to meet patron needs by metropolitan status rural suburban urban lt kbps gt kbps lt kbps gt kbps lt kbps gt kbps or fewer workstations . % n= . % n= . % n= . % n= . % n= . % n= more than workstations . % n= . % n= . % n= . % n= . % n= . % n= total . % n= . % n= . % n= . % n= . % n= . % n= table . public library connection speed insufficient to meet patron needs at some times by metropolitan status rural suburban urban lt kbps gt kbps lt kbps gt kbps lt kbps gt kbps or fewer workstations . % n= . % n= . % n= . % n= . % n= . % n= more than workstations . % n= . % n= . % n= . % n= . % n= . % n= total . % n= , . % n= . % n= . % n= . % n= . % n= table �. public library connection speed is sufficient to meet patron needs by metropolitan status rural suburban urban lt kbps gt kbps lt kbps gt kbps lt kbps gt kbps or fewer workstations . % n= , . % n= , . % n= . % n= , . % n= . % n= more than workstations . % n= . % n= . % n= . % n= . % n= . % n= total . % n= , . % n= , . % n= . % n= , . % n= . % n= , table . public library connection speed insufficient to meet patron needs some or all of the time by metropolitan status rural suburban urban lt kbps gt kbps lt kbps gt kbps lt kbps gt kbps or fewer workstations . % n= , . % n= . % n= . % n= . % n= . % n= more than workstations . % n= . % n= . % n= . % n= . % n= . % n= total . % n= , . % n= , . % n= . % n= , . % n= . % n= article title | author �assessing sufficiency and quality of bandwidth for public libraries | bertot and mcclure � ■ discussion and selected issues the data presented point to a number of issues related to the current state of public library pac and internet­access adequacy in terms of available public access computers and bandwidth. the data also provide a foundation upon which to discuss the nature of quality and sufficient pac and internet access in a public library environment. while public libraries indicate increased ability to meet patron bandwidth demand when providing fewer publicly avail­ able workstations, public libraries indicate that they have difficulty in meeting patron demand for public access computers. growth of wireless connections in , . percent of public library outlets offered wire­ less access, and a further . percent planned to make it available. outlets in urban and high­poverty areas were most likely to have wireless access. the majority of librar­ ies ( . percent), however, neither had wireless access nor had plans to implement it in . as table demon­ strates, the number of public library outlets offering wire­ less access has roughly doubled from . percent to . percent in two years. furthermore, . percent of outlets that do not currently have it plan to add wireless access in the next year. thus, if libraries follow through with their plans to add wireless access, . percent of public library outlets in the united states will have it by . the implications of the rapid growth of the public library’s provision of wireless connectivity (as shown in table ) on bandwidth requirements are significant. either libraries added wireless capabilities through their current overall bandwidth, or they obtained additional bandwidth to support the increased demand created by the service. if the former, then wireless access created an even greater burden on an already problematic band­ width capacity and may have actually reduced the overall quality of connectivity in the library. if the latter, libraries then had to shoulder the burden of increased expendi­ tures for bandwidth. either scenario required additional technology infrastructure, support, and expenditures. sufficient and quality connections the notion of sufficient and quality public library con­ nection to the internet is a moving target and depends on a range of factors and local conditions. for purposes of discussion in this paper, the authors used kbps to differentiate “slower” from “faster” connectivity. if, how­ ever, . mbps or greater had been used to define faster connectivity speeds, then only . percent of public libraries would meet the criterion of “faster” connectiv­ ity (see table ). and in fact, simply because . percent of public libraries report connection speeds of . mbps or faster does not also mean that they have sufficient or quality bandwidth to meet the computing needs of their users, their staff, their vendors, and their service provid­ ers. some public libraries may need mbps to meet the pac needs of their users as well as the internal staff and management computing needs. the library community needs to become more edu­ cated and knowledgeable about what constitutes sufficient and quality connectivity in their library for the communi­ ties that they serve. a first step is to understand clearly the nature and type of the connectivity of the library. the next step is to conduct an internal audit that minimally: ■ identifies the range of networked services the library provides both to users as well as for the operation of the library; ■ identifies the typical bandwidth consumption of these services; ■ determines the demands of users on the bandwidth in terms of services they use; ■ determines peak bandwidth­usage times; ■ identifies the impact of high­consumption networked services used at these peak­usage times; ■ anticipates bandwidth demands of newer services and resources that users will want to access through the library’s infrastructure—myspace.com, youtube. com—regardless of whether or not the library is the direct provider of such services; and ■ determines what broadband services are available to the library, the costs of these services, and the “fit” of these services to the needs of the library. based on this and related information from such an audit, library administration can better determine the degree to which the bandwidth is sufficient in speed and quality. ■ planning for sufficient and quality bandwidth knowing the current condition of existing bandwidth in the library is not the same as successful technology plan­ ning and management to ensure that the library has, in fact, bandwidth that is sufficient in speed and quality. once an audit such as has been suggested is completed, careful planning for bandwidth deployment in the library is essential. it appears, however, that currently much of the management and planning for networked services is based first on what bandwidth is available as opposed to the bandwidth that is needed to provide the necessary services and resources in a networked environment. this stance puts public libraries in a reactive condition rather than a proactive condition regarding provision of net­ worked services. information technology and libraries | march information technology and libraries | march most public library planning approaches stress the importance of conducting some type of needs assessment as a precursor to any type of planning. further, technology plans should include such things as goals, objectives, ser­ vices provision, and evaluation as they relate to bandwidth and the appropriate bandwidth needed. recent library technology planning guides, however, give little attention to the management, planning, and evaluation of band­ width as it relates to provision of networked services. it must be noted that some public libraries may be prevented from accessing higher bandwidth due to high cost, lack of availability of bandwidth alternatives, or other local factors that determine access to advanced telecommunications in their areas. in such circumstances, the audit may serve to inform the public service/utilities commissions, fcc, and others of the need for deploy­ ment of advanced telecommunications services in these areas. ■ bandwidth planning in a community context the audit and planning processes that have been described are critical activities for libraries. it is essential, however, for these processes to occur in the larger community con­ text. investments in technology infrastructure are increas­ ingly a community­wide resource that services multiple functions—emergency services, community access, local government agencies, to name a few. it is in this larger context that library pac and internet access occurs. moreover, there is a convergence of technology and service needs. for example, public libraries increasingly serve as agents of e­government and disaster­relief providers. first responders rely on the library’s infrastructure when theirs is destroyed, as hurricane katrina and other storms demonstrated. local, state, and federal government agen­ cies rely on broadband and pac and internet access (wired or wireless) to deliver e­government services. thus, at their core, libraries, emergency services, gov­ ernment agencies, and others have similar needs. pooling resources, planning jointly, and looking across needs may yield economies of scale, better service, and a more robust community technology infrastructure. emergency providers need access to reliable broadband and commu­ nications technologies in general, and in emergency situ­ ations in particular. libraries need access to high­quality broadband and pac technologies. both need access to wireless technologies. as broadcast networks relinquish ownership of the mhz frequency used for analog television in february , and this frequency is distributed to municipali­ ties for emergency services, now is an excellent time for libraries to engage in community technology planning for e­government, disaster planning and relief efforts, and pac and internet services. by working with the larger community to build a technology infrastructure, the library and the entire community benefit. ■ availability to high-speed connectivity one key consideration not known at this time is the extent to which public libraries—particularly those in rural areas—even have access to high­speed connec­ tions. many rural communities are served not by the large telecommunications carriers, but rather by small, privately owned­and­run local exchange carriers. iowa and wisconsin, for example, are each served by more than eighty exchange carriers. as such, public libraries are limited in capacity and services to what these exchange table . public-access wireless internet connectivity availability in public library outlets by metropolitan status and poverty metropolitan status poverty level provision of public-access wireless internet services urban suburban rural low medium high overall currently available . % ± . % (n= , ) . % ± . % (n= , ) . % ± . % (n= , ) . % ± . % (n= , ) . % ± . % (n= ) . % ± . % (n= ) . % ± . % (n= , ) not currently available and no plans to make it available within the next year . % ± . % (n= ) . % ± . % (n= , ) . % ± . % (n= , ) . % ± . % (n= , ) . % ± . % (n= , ) . % ± . % (n= ) . % ± . % (n= , ) not currently available, but there are plans to make it available within the next year . % ± . % (n= ) . % ± . % (n= , ) . % ± . % (n= , ) . % ± . % (n= , ) . % ± . % (n= ) . % ± . % (n= ) . % ± . % (n= , ) article title | author assessing sufficiency and quality of bandwidth for public libraries | bertot and mcclure carriers offer and make available. thus, in some areas, dsl service may be the only form of high­speed connec­ tivity available to libraries. and, as suggested earlier, dsl may or may not be considered high speed given the needs of the library and the demands of its users. communities that lack high­quality broadband ser­ vices by telecommunications carriers may want to con­ sider building a municipal wireless network that meets the community’s broadband needs for emergency, disas­ ter, and public­access settings. as a community engages in community­wide technology planning, it may become evident that local telecommunications carriers do not meet the broadband needs of the community. such com­ munities may need to build their own networks, based on identified technology­plan needs. ■ knowledge of networked services connectivity needs patrons may not attempt to use high­bandwidth services at the public library because they know from previous visits that the library cannot provide acceptable connec­ tivity speeds to access that service—thus, they quit trying to access that service, limiting the usefulness of the pub­ lic library. in addition, librarians may have inadequate knowledge or information to determine when bandwidth is or is not sufficient to meet the demands of their users. indeed, the survey and site visits revealed that some librarians did not know the connection speeds that linked their library to the internet. consequently, libraries are in a dilemma: increase both the number of workstations and the bandwidth to meet demand; or provide less service in order to operate within the constraints of current connectivity infrastruc­ ture. and yet, roughly percent of public libraries indi­ cate that they have no plans to add workstations within the next two years; the average number of workstations has been around ten for the last three surveys ( , , and ); and percent of public libraries indicate that space limitations affect their ability to add workstations. hence, for many libraries, adding workstations is not an option. ■ missing the mark? the networked environment is such that there are multi­ ple uses of bandwidth within the same library—for exam­ ple, public internet access, staff access, wireless access, integrated library system access. we are now in the web . environment, which is an interactive web that allows for content uploading by users (e.g., blogs, mytube.com, myspace.com, gaming). streaming content, not text, is increasingly the norm. there are portable devices that allow for text, video, and voice messaging. increasingly, users desire and prefer wireless services. this is a new environment in which libraries provide public access to networked services and resources. it is an enabling environment that puts users fully in the content seat—from creation to design to organization to access to consumption. and users have choices, of which the public library is only one, regarding the information they choose to access. it is an environment of competition, advanced applications, bandwidth intensity, and high­quality com­ puters necessary to access the graphically intense content. the impacts of this new and substantially more com­ plex environment on libraries are potentially significant. as user expectations rise, combined with the provision of high­quality services by other providers, libraries are in a competitive and service­ and resource­rich informa­ tion environment. providing “bare minimum” pac and internet access can have two detrimental effects in that they: ( ) relegate libraries to places of last resort, and ( ) further digitally divide those who only have public­access computers and internet access through their public librar­ ies. it is critical, therefore, for libraries to chart a high­end course regarding pac and internet access, and not access that is merely perceived to be acceptable by the librarians. ■ additional research the context in which issues regarding quality pac and sufficient connectivity speeds to internet access reside is complex and rapidly changing. research questions to explore include: ■ is it possible to define quality pac and internet access in a public library context? ■ if so, what are the attributes included in the defini­ tion? ■ can these attributes be operationalized and mea­ sured? ■ assuming measurable results, what strategies can the library, policy, research, and other interested communities employ to impact public library move­ ment toward quality pac and internet access? ■ should there be standards for sufficient connectivity and quality pac in public libraries? ■ how can public librarians be better informed regard­ ing the planning and deployment of sufficient and quality bandwidth? ■ what is the role of federal and state governments in supporting adequate bandwidth deployment for public libraries? ■ to what extent is broadband deployment and avail­ ability truly universal as per the universal service information technology and libraries | march information technology and libraries | march (section ) of the telecommunications act of (p.l. ­ )? these questions are a beginning point to a larger set of activities that need to occur in the research, practitioner, and policy­making communities. ■ obtaining sufficient and quality public-library bandwidth arbitrary connectivity speed targets, e.g., kbps or kbps, do not in and of themselves ensure quality pac and sufficient connectivity speeds. public libraries are indeed connected to the internet and do provide public­ access services and resources. it is time to move beyond connectivity­type and ­speed questions and consider issues of bandwidth sufficiency, quality, and the range of networked services that should be available to the public from public libraries. given the widespread connectivity now provided from most public libraries, there continue to be increased demands for more and better networked services. these demands come from governments that expect public libraries to support a range of e­government services, from residents who want to use free wireless connectivity from the public library, to patrons who need to download music or view streaming videos (to name but a few). simply providing more or better connectivity will not, in and of itself, address all of these diverse service needs. increasingly, pac support will require additional public librarian knowledge, resources, and services. sufficient and quality bandwidth is a key component of those services. the degree to which public libraries can provide such enhanced networked services (requiring exceptionally high bandwidth that is both sufficient and of high quality) is unclear. mounting a significant effort now to better understand existing bandwidth use and plan for future needs and requirements in individual public libraries is essential. in today’s networked envi­ ronment, libraries must stay competitive in the provision of networked services. such will require sufficient and high­quality connectivity and bandwidth. ■ acknowledgements the authors gratefully acknowledge the support of the bill & melinda gates foundation and the american library association for support of the public libraries and the internet study. data from that study have been incorpo­ rated into this paper. references . information institute, public libraries and the internet (tal­ lahassee, fla.: information use management and policy insti­ tute, ). all studies conducted since are available at: http://www.ii.fsu.edu/plinternet (accessed march , ). . u.s. federal communications commission, high speed services for internet access: status as of december , (wash­ ington, d.c.: fcc, ), available at http://www.fcc.gov/ bureaus/common_carrier/reports/fcc­state_link/iad/ hspd .pdf (accessed mar. , ). . j. c. bertot et al., public libraries and the internet (tal­ lahassee, fla.: information use management and policy insti­ tute, forthcoming), available at http://www.ii.fsu.edu/plinternet (accessed mar. , ). . j. c. bertot et al., “drafted: i want you to deliver e­ government,” library journal , no. (aug. ): – . . c. r. mcclure et al., planning and role setting for public libraries: a manual of options and procedures (chicago: ala, ); e. himmel and w. j. wilson, planning for results: a public library transformation process (chicago, ala, ). . j. c. bertot et al., “drafted: i want you to deliver e­gov­ ernment.”; p. t. jaeger et al., “the policy implications of internet connectivity in public libraries,” government information quarterly , no. ( ): – . . j. c. bertot et al., public libraries and the internet . . jaeger et al., “the policy implications of internet connec­ tivity in public libraries.” checking out facebook.com | charnigo and barnett-ellis author name and second author checking out facebook.com | charnigo and barnett-ellis author name and second author author id box for column layout while the burgeoning trend in online social networks has gained much attention from the media, few studies in library science have yet to address the topic in depth. this article reports on a survey of academic librar- ians concerning their perspectives toward facebook.com, an online network for students. findings suggest that librarians are overwhelmingly aware of the “facebook phenomenon.” those who are most enthusiastic about the potential of online social networking suggested ideas for using facebook to promote library services and events. few individuals reported problems or distractions as a result of patrons accessing facebook in the library. when problems have arisen, strict regulation of access to the site seems unfavorable. while some librarians were excited about the possibilities of facebook, the majority surveyed appeared to consider facebook outside the purview of professional librarianship. d uring the fall of , librarians noticed something unusual going on in the houston cole library (hcl) at jacksonville state university (jsu). students were coming into the library in droves. patrons waited in lines with photos to use the public­access scan­ ner (a stack of discarded pictures quickly grew). library traffic was noticeably busier than usual and the computer lab was constantly full, as were the public­access termi­ nals. the hubbub seemed to center around one particular web site. once students found available computers, they were likely to stay glued to them for long stretches of time, mesmerized and lost in what was later determined to be none other than “facebook addiction.” this addic­ tion was all the more obvious the day the internet was down. withdrawal was severe. soon after the librarians noticed this curious behavior, an article in the chanticleer, the campus newspaper for jsu, dispelled the mystery surrounding the web­site brouhaha. a campus reporter broke the exciting news to the jsu community that “after months of waiting and requests from across the country, it’s finally here. jsu is officially on the facebook.” the library suddenly became a popular hangout for students in search of computers to access facebook. apparently jsu jumped on the bandwagon relatively late. the facebook craze had already spread throughout other colleges and universities since the web site was founded in february by mark zuckerberg, a former student at harvard university. the creators of facebook vaguely define the site as “a social utility that connects you with the people around you.” although originally created to allow students to search for other students at colleges and universities, the site has expanded to allow individuals to connect in high schools, companies, and within regions. recently, zuckerberg has also announced plans to expand the network to military bases. currently, students and alumni in more than , colleges and uni­ versities communicate, connect with other students, and catch up with past high school classmates daily through the network. students who may never physically meet on campus (a rather serendipitous occurrence in nature) have the opportunity to connect through facebook. establishing virtual identities by creating profiles on the site, students post photographs, descriptions of academic and personal interests such as academic majors, campus organizations of which they are members, political orientation, favorite authors and musicians, and any other information they wish to share about themselves. facebook’s search engine allows users to search for students, faculty, and staff with similar interests by keyword. it would be hard to gauge how many of these students actually meet in person after connecting through facebook. the authors of this study have heard students mention that either they or their friends have made dates with other students on campus through facebook. many of the “friends” facebook users first add when they initially establish their accounts are the ones they are already acquainted with in the physical world. when facebook made its debut at jsu, it had become the “ninth most highly trafficked web site in the u.s.” one source estimated that percent of college students whose institutions are registered in facebook’s directory have created personal profiles on the site. membership for the university network requires a university e­mail address, and an institution cannot be registered in the directory unless a significant number of students request that the school be added. currently, more than nine mil­ lion people are registered on facebook. soon after jsu was registered on facebook’s direc­ tory, librarians began to receive questions regarding use of the scanner and requests for help uploading pictures to facebook profiles. students seemed surprisingly open about showing librarians their profiles, which usually contained more information than the librarians wanted to know. however, not all students were enthusiastic about facebook. complaints began to surface from students awaiting access to computers for academic work while classmates “tied up” computers on facebook. some stu­ dents complained about the distraction facebook caused checking out facebook.com: the impact of a digital trend on academic libraries laurie charnigo and paula barnett-ellis laurie charnigo (charnigo@jsu.edu) is an education librarian and paula barnett-ellis (pbarnett@jsu.edu) is a health, science, and nursing librarian at the houston cole library, jacksonville state university, alabama. information technology and libraries | march information technology and libraries | march in the library’s computer lab, a complaint that eventually reached the president of jsu. currently, the administra­ tion at jsu has decided to block access to facebook in the computer labs on campus, including the lab in the library. opinions of faculty and staff in the library about facebook vary. some librarians scoff at this new trend, viewing the site primarily as just another dating service. others have created their own facebook accounts just to see how it works, to connect with students, and to keep up with the latest internet fad. ■ study rationale prompted by the issues that have arisen at hcl as a result of heavy patron use of facebook, the authors surveyed academic librarians throughout the united states to find out what impact, if any, the site has had on other libraries. the authors sought information about the practical effect facebook has had on libraries, as well as librarians’ perspectives, perceived roles associated with, and awareness of internet social trends and their place in the library. online social networking, like e­mail and instant messaging, is emerging as a new method of com­ munication. recently, the librarians have heard facebook being used as a verb (e.g., “i’ll facebook you”). few would probably disagree that making social connections and friends (and facebook revolves around connecting friends) is an important aspect of the campus experi­ ence. much of the attraction students and alumni have toward college yearbooks (housed in the library) stems from the same fascination that viewing photos, student profiles, and searching for past and present classmates on facebook inspires. emphasis in this study centers on librarians’ awareness of, experimentation with, and atti­ tudes towards facebook and whether or not they have created policies to regulate or block access to the site on public­access computers. however trendy an individual web site such as facebook may appear, online social networking, a cat­ egory facebook falls within, has become a new subject of inquiry to marketing professionals, sociologists, commu­ nication scholars, and library and information scientists. downes defines social networks as a “collection of indi­ viduals linked together by a set of relations.” according to downes, “social networking web sites fostering the development of explicit ties between individuals as ‘friends’ began to appear in .” facebook is just one of many popular online social network sites (myspace, friendster, flickr), and survey respondents often asked why questions focused solely on facebook. the authors decided to investigate it specifically because it is cur­ rently the largest online social network targeted for the academic environment. librarians are also increasingly exploring the use of what have loosely been referred to as “internet . ” com­ panies and services, such as facebook, to interact with and reach out to our users in new and creative ways. the term internet . was coined by o’reilly media to refer to internet services such as blogs, wikis, online social net­ working sites, and types of networks that allow users the ability to interact and provide feedback. o’reilly lists the core competencies that define internet . services. one of these competencies, which might be of particular inter­ est to librarians, is that internet . services must “trust the users” as “co­developers.” as librarians struggle to develop innovative ways to reach users beyond library walls, it seems logical to observe online services, such as facebook and myspace, which appeal to a huge portion of our clientele. from a purely evaluative standpoint of the site as a database, the authors were impressed by several of the search features offered in facebook. graph­theory algo­ rithms and other advanced network technology are used to process connections. some of the more interesting search options available in facebook include the ability to: ■ search for students by course field, class number, or section; ■ search for students in a particular major; ■ search for students in a particular student organiza­ tion or club; ■ create “groups” for student organizations, clubs, or other students with common interests; ■ post announcements about campus or organization events; ■ search specifically for alumni; and ■ block or limit who may view profiles, providing users with built­in privacy protection if the user so wishes. since the authors finished the study, the site has added a news feed and a mini feed, features that allow users to keep track of their friends’ notes, messages, profile changes, friend connections, and group events. in response to negative feedback about the news feeds and mini feeds by users who felt their privacy was being violated, facebook’s administrators created a way for users to turn off or limit information displayed in the feeds. the addition of this technology, however, provides a sophisticated level of connectivity that is a benefit to users who like to keep abreast of the latest happenings in their network of friends and groups. the pulse, another feature on the site, keeps daily track of popular interests (e.g., favorite books) and member demographics (number of members, political orientation) and compares them with overall facebook member averages. the authors were pleasantly surprised to discover that the beatles and led zeppelin, beloved bands of the baby boomers, article title | author checking out facebook.com | charnigo and barnett-ellis continue to live on in the hearts of today’s students. these groups were ranked in the top ten favorite bands by stu­ dents at jsu. as of october , the top campaign issues expressed by facebook users were: reducing the drinking age to eighteen (go figure) and legalization for same­sex marriage. arguably, much of the information provided by facebook is not academic in nature. however, an evaluation or review of facebook might provide useful information to instruction librarians and database ven­ dors regarding interface design and search capabilities that appeal to students. provitera­mcglynn suggests that facilitating learning among millennials, who “represent to million people” born after (a large percent­ age of facebook members) involves understanding how they interact and communicate. awareness of students’ cultural and social interests, and how they interact online, may help older generations of academic librarians better connect with their constituents. ■ the literature on online social networks although social networks have been the subject of study by sociologists for years and social network theories have been established to describe how these networks func­ tion, the study of online social networks has received little attention from the scholarly community. in , garton, haythornthwaite, and wellman were among the first to describe a method, social network analysis, for studying online social networks. their work was published years before online social networks similar to facebook evolved. currently, the literature on these networks is predominantly limited to popular news pub­ lications, business magazines, occasional blurbs in library science and communications journals, and numerous student newspapers. privacy issues and concerns about sexual predators lurking on facebook and similar sites have been the focus of most articles. in the chronicle of higher education, read details numerous arrests, suspensions, and schol­ arship withdrawals that have resulted from police and administrators searching for incriminating information students have posted in facebook. read discovered that, because students naively reveal so much informa­ tion about their activities, some campus police were regularly trolling facebook, finding it “an invaluable ally in protecting their campuses.” students may feel a false sense of security when they post to facebook, regarding it as their private space. however, read warns that “as more and more colleges offer alumni e­mail accounts, and as campus administrators demonstrate more internet savvy, students are finding that their conversations are playing to a wider audience than they may have antici­ pated.” privacy concerns expressed about facebook appear to revolve more around surveillance than stalk­ ers. in a web seminar on issues regarding facebook use in higher education, shawn mcguirk, director of judicial affairs, mediation, and education at fitchburg state college, massachusetts, recommends that administrators and others concerned with students posting potentially incriminating, embarrassing, or overtly personal infor­ mation draft a document similar to the one created by cornell university’s office of information technologies, which advises students on how to safely and responsibly use online social networking sites similar to facebook. after pointing out the positive benefits of facebook and reassuring students that cornell university is proud of its liberal policy in not monitoring online social networks, the essay, entitled “thoughts on facebook,” provides poignant advice and examples of privacy issues revolv­ ing around facebook and similar web sites. the golden rule of this essay states: don’t say anything about someone else that you would not want said about yourself. and be gentle with your­ self too! what might seem fun or spontaneous at , given caching technologies, might prove to be a liability to an on­going sense of your identity over the longer course of history. a serious concern discussed in this document is the real possibility that potential employers may scan facebook profiles for the “real skinny” on job candidates. however, unless the employer uses an e­mail issued from the same school as the candidate, he or she is unable to look at the individual’s full profile without first request­ ing permission from the candidate to be added as a “friend.” all the employer is able to view is the user’s name, school affiliation, and picture (if the user has posted one). unless the user has posted an inappropriate picture or is applying for a job at the college he or she is attending, the threat of employers snooping for informa­ tion on potential candidates in facebook is minimal. the same, however, cannot be said of myspace, which is much more open and accessible to the public. additionally, three pilot research studies have also focused on privacy issues specifically relating to facebook, including those of stutzman, gross and acquisti, and govani and pashley. results from all three studies revealed strikingly close findings. individuals who participated in the studies seemed willing to dis­ close personal information about themselves—such as photos and sometimes even phone numbers and mailing addresses—on facebook profiles even though students also seemed to be aware that this information was not secure. in a study of fifty carnegie mellon university undergraduate users, govani and pashley concluded that these users “generally feel comfortable sharing their per­ sonal information in a campus environment. participants said they “had nothing to hide” and “they don’t really information technology and libraries | march information technology and libraries | march care if other people see their information.” a separate study of more than four thousand facebook members at the same institution by gross and acquisti echoed these findings. comparing identity elements shared by members of facebook, myspace, friendster, and the university of north carolina directory, stutzman discov­ ered that a significant number of users shared personal information about themselves in online social networks, particularly facebook, which had the highest level of campus participation. gross and acquisti provide a list of explanations suggesting why facebook members are so open about sharing personal information online. three explanations that are particularly convincing are that “the perceived benefit of selectively revealing data to strang­ ers may appear larger than the perceived costs of possible privacy invasions”; “relaxed attitudes toward (or lack of interest in) personal privacy”; and “faith in the network­ ing service or trust in its members.” in public libraries, concern has primarily centered on teenagers accessing myspace.com, an online social net­ working site much larger than facebook. myspace, whose membership, unlike facebook, does not require an .edu e­mail address, has a staggering million users, a num­ ber that continues to rise. julian aiken, a reference librar­ ian at the new haven free public library, wrote about the unpopular stance he took when his library decided to ban access to myspace due to the hysterical hype of media reports exposing the dangers from online predators lurking on the site. for aiken, the damage of censorship policies in libraries far outweighs the potential risk of sex crimes. furthermore, he suggests that there are even edu­ cational benefits of myspace, observing that “[t]eenagers are using myspace to work on collaborative projects and learn the computer and design skills that are increasingly necessary today.” what is apparent is that whether facebook continues to rise in popularity or fizzles out among the college crowd, the next generation of college students, who now constitute the largest percentage of myspace users, are already solidly entrenched and adept at using online social networks. librarians in institutions of higher education might need to consider what implica­ tions the communication style preferences of these future students could have, if any, on library services. while most of the academic attention regarding online social networks has centered on privacy concerns, perhaps the business sector has done a more thorough investiga­ tion of user behavior and students’ growing attraction towards these types of sites. business magazines have naturally focused on the market potential, growth, and fluctuating popularity of various online social networks. advertisers and investors have sought ways to capital­ ize on the exponential growth of these high­traffic sites. business week reported that as of october , facebook .com had . million members. more than half of those members were between the ages of twelve and twenty­ four. while some portended that the site was losing momentum, as of august , membership on facebook had expanded beyond eight million. marketing experts have closely studied, apparently more so than com­ munication scholars, the behavior of users in online social networks. in a popular business magazine, hempel and lehman describe user behavior of the “myspace generation”: “although networks are still in their infancy, experts think they’re already creating new forms of social behavior that blur the distinctions between online and real­world interactions.” the study of user behavior in online social networks, however, has yet to be addressed in length by those outside the field of marketing. although evidence of interest in online social net­ works is apparent in librarian weblogs and forums (many librarians have created facebook groups for their libraries), actual literature in the field of library and information science is scarce. dvorak questions the lack of interest displayed by the academic community toward online social networks as a focus of scholarly research. calling on academics to “get to work,” he argues “aca­ demia, which should be studying these phenomena, is just as out of the loop as anyone over .” this discon­ nect is also echoed by michael j. bugeja, director of the greenlee school of journalism and communication at iowa state university, who writes, “while i’d venture to say that most students on any campus are regular visitors to facebook, many professors and administrators have yet to hear about facebook, let alone evaluate its impact.” the lack of published research articles on these types of networks, however, is understandable given the newness of the technology. a few members of the academic community have sug­ gested opportunities for using facebook to communicate with and reach out to students. in a journal specifically geared toward student services in higher education, shier considers the impact of facebook on campus community building. although she cannot identify an academic purpose for facebook, she describes how the site can con­ tribute to the academic social life of a campus. facebook provides students with a virtual campus experience, particularly in colleges where students are commuters or are in distance education. shier writes, “as the student’s definition of community moves beyond the geographic and physical limitations, facebook.com provides one way for students to find others with common interests, feel as though they are part of a large community, and also find out about others in their classes.” furthermore, facebook membership extends beyond students to fac­ ulty, staff, and alumni. shier cites examples of professors who used facebook to connect or communicate with their students, including the president of the university of iowa and more than one hundred professors at duke university. professors who teach online courses make article title | author checking out facebook.com | charnigo and barnett-ellis themselves seem more human or approachable by estab­ lishing facebook profiles. greeting students on their own turf is exactly the direction staff at washington university’s john m. olin library decided to take when they hired web services librarian joy weese moll to communicate and answer questions through a variety of new technologies, includ­ ing facebook. brian mathews, information services librarian at georgia institute of technology, also created a facebook profile in order to “interact with the students in their natural environment.” mathews decided to experiment with the possibilities of using facebook as an outreach tool to promote library services to , stu­ dents in the school of mechanical engineering after he discovered that , of these students were registered on facebook. advising librarians to become proactive in the use of online social networks, mathews reported that overall, his experience helped him to effectively “expand the goal of promoting the library.” bill drew was among the first librarians to create an account and profile for his library, the suny morrisville library. as of september , nearly one hundred librarians had created profiles or accounts for their libraries on facebook. one month later, however, the administration at facebook began shutting down library accounts on the grounds that libraries and institutions were not allowed to represent themselves with profiles as though they were individu­ als. in response, many of these libraries simply created groups for their libraries, which is completely appropri­ ate, similar to creating a profile, and just as searchable as having an account. the authors of this study created the “houston cole library users want answers!” group, which currently has ninety­one members. library news and information of interest about the library is announced in the group. in this study, one trend the authors will try to identify is whether other librarians have considered or are already using facebook in similar ways that moll, mathews, and drew have explored as avenues for com­ municating with students or promoting library services. ■ the survey in february , surveys were mailed to reference or public service librarians (when the identity of those per­ sons could be determined). these individuals were chosen from a random sample of the institutions of higher education classified by the carnegie classification listing of higher education institutions as “master’s colleges and universities (i and ii)” and “doctoral/ research universities (extensive and intensive).” the sample size provided a . percent margin error and a percent confidence level. one hundred twenty­six surveys were completed, providing a response rate of percent. fifteen survey questions (appendix a) were designed to target three areas of inquiry: awareness of facebook, practical impact of the site on library services, and perspectives of librarians toward online social networks. awareness of facebook a series of questions on the survey queried respondents about their awareness and degree of knowledge about facebook. the overwhelming majority of librarians were aware of facebook’s existence. out of librarians, had at least heard of facebook; were not familiar with the site. as one individual wrote, “i had not heard of facebook before your survey came, but i checked and our institution is represented in facebook.” universities registered in facebook are easily located through a search­by­region on facebook’s home page. thirty­eight colleges and universities for alabama (jsu’s location) are registered in facebook. (in comparison, academic institutions in california are listed.) out of those librar­ ians who had heard of the site, were not sure whether their institutions were registered in facebook’s directory. sixty survey participants were aware that their institu­ tions were registered in the directory, while fifteen librar­ ians reported that their universities were not registered (figure ). several comments at the end of the survey indicated that some of the institutions surveyed did not issue school e­mail accounts, making membership in facebook impossible for their university. interestingly, out of the sixty individuals who could claim that their universities were in the directory, percent have created their own personal facebook accounts and two libraries have individual profiles (figure ). one individual who established an account on the site wrote, “personally, i’m a little embarrassed by having an account because it’s such a teeny­bopper kind of thing and i’m a little old for it. but it’s an interesting cultural phenomenon and academic librarians need to get on the bandwagon with it, if only to better understand their constituents.” another survey respondent with an individual profile on the site reported a group created by his or her institution on facebook titled “i totally want to have sex in the library.” this individual wanted to make it clear, however, that the students—not the librarians—created this group. a particularly help­ ful participant went so far as to poll the reference col­ leagues in all nine of the libraries at his/her institution and found that “only a few had even heard of facebook.” that librarians will become increasingly aware of online social networks was the sentiment expressed by another individual who wrote, “most librarians at my institu­ tion are unaware of social software in general, much less facebook. however, i think this will change in the future as social software is mentioned more often in traditional media (such as television and newspapers).” according to survey responses, it does not appear information technology and libraries | march information technology and libraries | march that use of facebook by students has been as noticeable or distracting in other libraries as it has been at hcl. when asked to describe their observation of student use of library computers to access facebook, percent of those surveyed checked “rarely to never.” only percent indicated “most of the time” to “all of the time” (table ). however, it is important to remember that only sixty individuals could verify that their institutions are regis­ tered on facebook. through comments, some librarians hinted that “snooping” or keeping mental notes of what students view on library computers is frowned upon. it simply is not our business. “we do not regulate or track student use of computers in the library,” wrote one indi­ vidual. several librarians noted that students were using facebook in the libraries, but more so on personal laptops than public­access computers. practical impact of facebook another goal of this study was to find out whether facebook has had any real impact on library services, such as an increase in bandwidth, library traffic, and noise, or in use of public­access computers, scanners, or other equipment. student complaints about monopolization of computers for use of facebook led administrators to block the site from computer labs at jsu. access to facebook on public­access terminals, however, was not regulated. survey responses revealed that facebook has had minimal impact on library services elsewhere. only one library was forced to develop a policy for specifically addressing computer­use concerns as a result of facebook use. one individual mailed the sign posted on every computer terminal in the library, which states, “if you are using a computer for games, chat, or other recreational activity, please limit your usage to thirty minutes. computers are primarily intended for academic use.” another librarian reported that academic computing staff had to shut down access to facebook on library computers due to band­ width and access issues. this individual, however, added, “interestingly, no one has complained to the library staff about its absence!” given a list of possible effects facebook may have had on library services and operations, per­ cent of respondents indicated that facebook has increased patron use of computers. seven percent agreed that it has increased patron traffic, and only percent reported that the site has created bandwidth problems or slowed down internet access. only four individuals received patron complaints about other users “tying up” the computers with facebook (figure ). since the advent of facebook, the public scanner has become one of the hottest items in hcl. librarians at jsu know that use of the scanner has increased tremendously due to facebook because the scanner used by students to upload photos is attached to a public workstation next to the general reference desk. students often ask questions about uploading pictures to their facebook profiles as well as how to edit photos (e.g., resizing and cropping). one survey question asked whether scanner use had increased as a result of facebook. of the sixty­two respon­ dents who answered this question (it was indicated that only those libraries that provide public access to scanners should answer the question), percent reported that figure . institutions added to the facebook directory figure . involvement with facebook table . student use of library computers to access facebook (based on observation) total percentage never rarely some of the time all the time most of the time article title | author �checking out facebook.com | charnigo and barnett-ellis � scanner use had not increased. furthermore, only two librarians have assisted students with the scanner or pro­ vided any other type of assistance, for that matter, with facebook. the assistance the two librarians gave included scanning photographs, editing photos, uploading photos to facebook profiles, and creating accounts. however, in a separate question, percent of participants agreed that librarians should be responsible for helping students, when needed, with questions about facebook. no librar­ ian has added additional equipment such as computers or scanners as a result of facebook. only one individual reported future plans by his/her library to add additional equipment in the future as a result of heavy use of the site. perspectives toward facebook one of the main goals of the study was to obtain a snapshot of the perspectives and attitudes of librarians toward facebook and online social networks in general. most of the librarians surveyed were neither enthusiastic nor disdainful of facebook. a small group of the respon­ dents, however, when given the chance to comment, were extremely positive and excited about the possibilities of online social networking. twenty­one individuals saw no connection between libraries and facebook. sixty­ seven librarians were in agreement that computer use for academic purposes should take priority, when needed, over use of facebook. however, fifty­one respondents indicated that librarians needed to keep up with internet trends, such as facebook, even when such trends are not academic in nature (table ). out of librarians who completed the survey, only reported that facebook has generated discussion among library faculty and staff about online social networks. on the other hand, few individuals voiced negative opinions toward facebook. only percent of those surveyed indicated that facebook annoyed faculty and staff. one individual wrote, “i don’t like facebook or most social networking services. they encourage the formation of cliques and keep users from meeting and accepting those who are different than themselves.” comments like this, however, were rare. although the majority of librarians seemed fairly apa­ thetic toward facebook, few individuals expressed nega­ tive comments toward the site. few librarians indicated that facebook should be addressed or regulated in library policy. most individu­ als viewed the site as just another communication tool similar to instant messaging or cell phones. in fact, while most librarians did not express much interest in facebook, many were quite vocal about not regulating its use. the following comment by one survey partici­ pant captures this sentiment: “attempts to restrict use of facebook in the library would be futile, in my opinion, in the same way it is now impossible to ban use of usb drives and aim in academic libraries.” while most indi­ table . access, assistance, and awareness of facebook and similar trends: perspectives total percentage computer use for academic purposes should take priority, when needed, over use of facebook. librarians need to “keep up” with internet trends, such as facebook, even when these trends are not aca- demic in nature. library resources should not be monopolized with use of facebook. librarians should help students, when able, with questions regarding facebook. there is no connection between libraries and facebook. student use of facebook on library computers should not be regulated. library computers should be available for access to facebook, but librarians should not feel that it is their responsibility to assist students with questions regarding the site. (respondents were allowed to check any or all responses that applied.) figure . patron complaints about facebook information technology and libraries | march information technology and libraries | march viduals agreed that academic use of computers should take priority over recreational use, a polite request that a patron using facebook allow another student to use the computer for academic purposes, when necessary, appears more preferable than the creation and enforce­ ment of strict policies. as one librarian put it, “i don’t want students to see the library as a place where they are ‘policed’ unnecessarily.” when asked if facebook serves any academic pur­ pose, percent of those surveyed indicated that it does not, while percent were “not sure.” twelve percent of the librarians identified academic potential or pos­ sible benefits of the site (figure ). the authors were surprised to find that percent of those surveyed were not completely willing to dismiss facebook as pure rec­ reation. some librarians found facebook to be a distrac­ tion to academics: “maybe i’m old fashioned, but when do students find time for this kind of thing? i wonder about the impact of distractions like this on academic pursuits. there’s still only twenty­four hours in a day.” another individual asked two students who were using facebook in the library what they thought of the site and they admitted that it was “frequently a distraction from academic work.” for the percent who were not sure whether facebook has any academic value, there were comments such as “i am continuing to observe and will decide in the future.” academic uses for facebook included suggestions that it be used as a communication tool for student collaboration in classes (facebook allows students to search for other students by course and sec­ tion number). one individual suggested it could be used as an “online study hall,” but then wondered if this might lead to plagiarism. some thought instructors could somehow use facebook for conducting online discussion forums, with one participant observing “it’s ‘cooler’ than using blackboard.” “building rapport” with students through a communication medium that many students are comfortable with was another benefit mentioned. respondents who were enthusiastic about facebook thought it most beneficial as a virtual extension of the campus. facebook could potentially fill a void where face­to­face connections are absent in online and dis­ tance­education classes. several librarians suggested that facebook has had a positive influence in fostering col­ legiate bonds and school spirit. as one individual wrote, “[t]he academic environment is not only responsible for scholarly growth, but personal growth as well. this is just one method for students to interact in our highly techno­ logical society.” facebook could provide students who are not physically on campus with a means to connect with other students at their institutions who have similar academic and social interests. some librarians were so enthusiastic about facebook that they suggested libraries use the site to promote their services. using the site to advertise library events and creating online library study groups and book clubs for students were some of the ideas expressed. one librar­ ian wrote: “facebook (and other social networking sites) can be a way for libraries to market themselves. i haven’t seen students using facebook in an academic manner, but there was a time when librarians frowned on e­mail and aim too. if it becomes a part of students’ lives, we need to welcome it. it’s part of welcoming them, too.” more librarians, however, felt that facebook should serve as a space exclusively for students and that librarians, profes­ sors, administrators, police, and other uninvited folks should keep out. furthermore, as one individual noted, it is not “an appropriate venue” for librarians to promote their services. while the review of literature demonstrates that much has been made of online social networks and privacy issues, the librarians surveyed were not particularly con­ cerned about privacy. only percent indicated that they were concerned about privacy issues related to facebook. however, some librarians voiced concerns that many stu­ dents are ignorant about the risks of posting personal infor­ mation and photographs on facebook and do not seem fully aware of the possibility that individuals outside their social sphere might also have reason to access the site. one individual mentioned that the librarians at her institution have begun to emphasize this to students during library instruction sessions on internet research and evaluation. ■ limitations several limitations to this study must be noted when attempting to reach any type of conclusion. participants who had never heard of facebook obviously could not answer any questions except that they were not famil­ iar with the site. some questions required respondents to “guesstimate.” unless librarians have access to their figure . finds conceivable academic value in facebook article title | author checking out facebook.com | charnigo and barnett-ellis institution’s internet usage statistics, it would be hard for them to really know how much bandwidth is being used by students accessing facebook. librarians, having been trained in a profession that places a high value on freedom of access, might also be wary of activities that suggest any type of censorship. therefore, it is conceivable that some of the librarians surveyed do not know whether students are using facebook in the library because they make a point not to snoop or make note of individual web sites that students view. ■ discussion while online education is growing at a rapid rate across the united states, so is the presence of virtual academic social communities. although facebook might prove to be a passing fad, it is one of the earliest and largest online social networking communities geared specifically for students in higher education. it represents a new form of communication that connects students socially in an online environment. if online academics have evolved and continue to do so, then it is only natural that online academic social environments, such as facebook, will continue to evolve as well. while traditionally considered the heart of the campus, one is left to ponder the library’s presence in online academic social networks. what role the library will serve in these environments might largely depend on whether librarians are proactive and experi­ mental with this type of technology or whether they simply dismiss it as pure recreation. emerging technolo­ gies for communication should provoke, at the very least, an interest in and knowledge of their presence among library and information science professionals. this survey found that librarians were overwhelmingly aware of and moderately knowledgeable about facebook. some librarians were interested in and fascinated with facebook, but preferred to study it as outsiders. others had adopted the technology, but more for the purpose, it would seem, of having a better understanding of today’s students and why facebook (and other online social net­ working sites) appeals to so many of them. it is apparent from this study that there is a fine line between what now constitutes “academic” activity and “recreational” activity in the library. sites like facebook seem to blur this line fur­ ther and librarians do not seem eager or find it necessary to distinguish between the two unless absolutely pressed (e.g., asking a student to sign out of facebook when other patrons are waiting to use computers for academic work). one area of attention this study points to is a lack of con­ cern among librarians toward the internet and privacy issues. some individuals surveyed suggested that librari­ ans play a larger role in making students aware that people outside their society of friends—namely, administrative or authority figures—have the ability to access the informa­ tion they post online to social networks. participants were most enthusiastic about facebook’s role as a space where students in the same institution can connect and share a common collegiate bond. librarians who have not yet “checked out” facebook might consider one individual’s description of the site as “just another ver­ sion of the college yearbook that has become interactive.” among the most cherished books in hcl that document campus life at jsu are the mimosa yearbooks. alumni and students regularly flip through this treasure trove of pho­ tographs and memories. no administrator or librarian would dare weed this collection or find its presence irrele­ vant. while year books archive campus yesteryears, online social networks are dynamically documenting the here and now of campus life and shaping the future of how we communicate. as casey writes, “libraries are in the habit of providing the same services and the same programs to the same groups. we grow comfortable with our provision and we fail to change.” by exploring popular new types of internet services such as facebook instead of quickly dismissing them as irrelevant to librarianship, we might learn new ways to reach out and communicate better with a larger segment of our users. ■ acknowledgements the authors would like to acknowledge stephanie m. purcell, student worker at the houston cole library, for her excellent editing suggestions and insight into online social networks from the student’s point of view, and john­bauer graham, head of public services at the houston cole library, for his encouragement. references and notes . angela reid, “finally . . . the facebook,” the chanticleer, sept. , , . . facebook.com, http://www.facebook.com/about.php (accessed dec. , ). . angus loten, “the great communicator,” inc.com., june , , http://www.inc.com/ under /zuckerberg.html (accessed dec. , ). . adam lashinsky, “facebook stares down success,” for- tune, nov. , , . . michael amington, “ percent of college students use facebook,” testcrunch: tracking web . company review on face- book (sept. , ), http://www.techcrunch.com/ / / / ­of­college­students­use­facebook (accessed dec. , ). . http://www.facebook.com/about.php. . facebook us! if you are a registered member of facebook, do a global search for “laurie charnigo” or “paula barnett­ ellis.” information technology and libraries | march information technology and libraries | march . stephen downes, “semantic networks and social net­ works,” the learning organization , no. ( ): . . ibid. . tim o’reilly, “what is web . ?” http://www.oreilly net.com/pub/a/oreilly/tim/news/ / / /what­is­web ­ .html (accessed aug. , ). . http://www.facebook.com/about.php. . angela provitera mcglynn, “teaching millennials, our newest cultural cohort,” the education digest , no. ( ): . . laura garton, caroline haythornthwaite, and barry well­ man, “studying online social networks,” journal of computer mediated communication , no. ( ). . facebook.com’s “about” page archives a collection of col­ lege newspaper articles about facebook: http://www.facebook .com/about.php (accessed dec. , ). . brock read, “think before you share,” the chronicle of higher education, jan. , , a –a . . ibid., a . . ibid., a . . shawn mcguirk, “facebook on campus: understanding the issues,” magna web seminar presented live on june , . transcripts available for a fee from magna pubs. http://www .magnapubs.com/catalog/cds/ ­ .html (accessed aug. , ). . tracy mitrano, “thoughts on facebook” (apr. ) cor­ nell university of information technologies, http://www.cit .cornell.edu/oit/policy/ memos/facebook.html (accessed june , ). . ibid., “conclusion.” . tabreez govani and harriet pashley, “student awareness of the privacy implications when using facebook,” unpublished paper presented at the “privacy poster fair” at the carnegie mellon university school of library and information science, dec. , , , http://lorrie.cranor.org/courses/fa /tubzhlp .pdf (accessed jan. , ). . ralph gross and alessandro acquisti, “information rev­ elation and privacy in online social networks,” paper presenta­ tion at the acm workshop on privacy in the electronic society, alexandria, va., nov. , , , http://portal.acm.org/citation .cfm?id= (accessed nov. , ). . frederic stutzman, “an evaluation of identity­sharing behavior in social network communities,” paper presentation at the idmaa and ims code conference, oxford, ohio, april – , , – , http://www.ibiblio.org/fred/pubs/stutzman _pub .pdf (accessed may , ). . gross and acquisti, “information revelation and privacy in online social networks,” . . “myspace: design anarchy that works,” business week, jan. , , . . julian aiken, “hands off myspace,” american libraries , no. ( ): . . ibid. . jessi hempel and paula lehman, “the myspace genera­ tion,” business week, dec. , , . . http://www.facebook.com/about.php. . hempel and lehman, “the myspace generation,” . . the authors created the “librarians and facebook” group on facebook to discuss issues concerning facebook and librari­ anship, such as censorship issues, policies, and ideas for con­ necting with students through facebook. this is a global group. if you have a facebook account, we invite you to do a search for “librarians and facebook” and join our group. . john c. dvorak, “academics get to work!” pcmagazine online, http://www.pcmag.com/article / , , , .asp (accessed feb. , ). . michael j. bugeja, “facing the facebook,” the chronicle of higher education, jan. , , c –c ; ibid. . maria tess shier, “the way technology changes how we do what we do,” new directions for student services (winter ): – . . ibid., . shier, “the way technology changes how we do what we do,” ; j. duboff, “poke” your prof: faculty discovers thefacebook.com,” yale daily news, mar. , , http://www .yaledailynews.com/article.asp?aid= (accessed jan. , ; mingyang liu, “would you friend your professor? duke chronicle online, feb. , , http://www.dukechronicle.com/ media/paper /news/ / / /news/would.you.friend .your.professors­ .shtml?norewrite&sourcedomain =www.dukechronicle.com (accessed jan. , ). . brittany farb, “students can ‘check out’ new librarian on the facebook,” student life (washington univ. in st. louis), feb. , , http://www.studlife.com/home/index.cfm?eve nt=displayarticle&ustory_id= a d­ b (accessed feb. , ). . brian s. mathews, “do you facebook? networking with students online,” college & research libraries news , no. ( ): . . ibid., . . view the “houston cole library users want answers!” group by doing a search for the group title on facebook. . nces compare academic libraries, http://nces.ed.gov/ surveys/libraries/ compare/peervariable.asp (accessed dec. , ). the random sample was chosen using the research ran­ domizer available online, http://www.randomizer.org/form .htm (accessed dec. , ). . michael e. casey and laura c. savastinuk, “library . ,” library journal , no. ( ): . article title | author checking out facebook.com | charnigo and barnett-ellis . has your institution been added to the facebook directory?  yes  no (skip to questions , , and  not sure (skip to questions , , and )  i am not familiar with facebook (skip all questions and submit) . which best describes your involvement with facebook?  i have a personal account  my library has an account  no involvement . which best describes your observation of student use of library computers to access facebook?  all the time  most of the time  some of the time  rarely  never . has your library added additional equipment such as computers or scanners as a result of facebook use?  yes  no  no, but we plan to in the future . have patrons complained about other patrons using library computers for facebook?  yes  no  not sure . has your library had to develop a policy or had to address computer use concerns as a result of facebook use?  yes  no  not sure . if your library provides public access to a scanner, has patron use of scanners increased due to the use of facebook?  yes  no . have you assisted students with the library’s scan­ ner for facebook?  yes  no . if you have provided assistance to students with facebook, please check all that apply:  creating accounts  scanning photographs or offering advice on where students can access a scanner  editing photographs (e.g., resizing photos or use of a photo editor)  uploading photographs to facebook profiles  other __________________________________ . check the responses that best describe your opinion about the responsibilities of librarians in assisting students with facebook questions and access to the web site:  student use of facebook on library computers should not be regulated.  library resources should not be monopolized with facebook use.  computer use for academic purposes should take priority, when needed, over use of facebook.  librarians should help students, when able, with facebook questions.  librarians need to “keep up” with internet trends, such as facebook, even if they are not academic in nature.  there is no connection between librarians, libraries, and facebook.  library computers should be available for facebook use, but librarians should not feel that they need to assist students with facebook questions. . would you consider facebook to be a relevant aca­ demic endeavor?  yes  no  not sure appendix a: survey on the impact of facebook on academic libraries information technology and libraries | march information technology and libraries | march . if you answered “yes” to question , please describe how facebook could be considered an aca­ demic endeavor. ______________________________________________ ______________________________________________ ______________________________________________ ______________________________________________ . please check all answers that best describe what effect, if any, use of facebook in the library has had on library services and operations?  has increased patron traffic  has increased patron use of computers  has created computer access problems for patrons  has created bandwidth problems or slowed down internet access  has generated complaints from other patrons  annoys library faculty and staff  interests library faculty and staff  has generated discussion among library faculty and staff about facebook . is privacy a concern you have about students using facebook in the library?  yes  no  not sure please list any observations, concerns, or opinions you have regarding facebook use in libraries. extracted the paragraphs from my palm to my desktop, and saved that document and the tocs on a universal serial bus (usb) key. today, i combined them in a new document on my laptop and keyed the remaining paragraphs in my room at an inn on a pier jutting into commencement bay in tacoma on southern puget sound. i sought inspiration from the view out my window of the water and the fall color, from old crow medicine show on my ipod, and from early sixties beyond the fringe skits on my treo. fred kilgour was committed to delivering informa­ tion to users when and where they wanted it. libraries must solve that challenge today, and i am confident that we shall. editorial continued from page information technology and libraries | march information technology and libraries | march author id box for column layout opac design enhancements and their effects on circulation and resource sharing within the library consortium environment michael j. bennett a longitudinal study of three discrete online public access catalog (opac) design enhancements examined the pos- sible effects such changes may have on circulation and resource sharing within the automated library consor- tium environment. statistical comparisons were made of both circulation and interlibrary loan (ill) figures from the year before enhancement to the year after implemen- tation. data from sixteen libraries covering a seven-year period were studied in order to determine the degree to which patrons may or may not utilize increasingly broader opac ill options over time. results indicated that while ill totals increased significantly after each opac enhancement, such gains did not result in signifi- cant corresponding changes in total circulation. m ost previous studies of online public access catalog (opac) use and design have centered on transaction­log analysis and user survey results in the academic library environment. measures of patron success or lack thereof have traditionally been expressed in the form of such concepts as “zero­hit” analysis or the “branching” analysis of kantor and, later, ciliberti. missing from the majority of the literature on opac study, however, are the effects that use and design have had on public library patron borrowing practices. major drawbacks to transaction­log analyses and user surveys as a measure of successful opac use include a lack of standardization and the inherent difficulties in interpreting resulting data. as peters notes, “[s]urveys measure users’ opinions about online catalogs and their perceptions of their successes or failures when using them, while transaction logs simply record the searches conducted by users. surveys,” he concludes, “mea­ sure attitudes, while transaction logs measure a specific form of behavior.” in both cases it is difficult, in many instances, to draw clear conclusions from either method. circulation figures, on the other hand, measure a more narrowly defined level of patron success. circulation is a discrete output that is the direct result of patrons’ initiated interaction with one or many library collections, one or many levels of library technology. with the recent advent of such enhanced opac functionality as patron­placed holds on items from broader and broader catalogs, online catalogs now more than ever not only serve as search mechanisms but also as ways for patrons to directly obtain materials from multiple sources. it follows that an investigation of the possible effects such enhancements may have on general circulation trends is warranted. ■ literature review during the mid­to­late s, transaction­log analysis was introduced as an inexpensive and easy method of looking at opac use in primarily the academic library environment. peters’s transaction­log survey of more than thirteen thousand searches executed over a five­ month period at the university of missouri­kansas city remains particularly instructive today for its large sample and transferable design as well as its interpreta­ tion of results. here analysis was broken into two phases. in phase one, usage patterns by search type and failure rates as measured by zero hits were examined as dependent vari­ ables with search type as the independent variable in a comparison study. phase two took this one step further in the assigning of what peters termed “probable cause” of zero hits. these probable causes fell into patterns that, in turn, resulted in the identification of fourteen discernable error types that included such things as typographical errors and searches for items not in the catalog. once again, search type formed the independent variable while error type shaped the dependent variable in a simple study of error types as a percentage of total searches. peters found that users rarely employed truncation or any advanced feature searches and that failures were due primarily to such consistent erroneous search patterns as typographical errors and misspellings. more importantly, however, he cogently reassessed transaction­log analysis as a tool and critiqued its limitations. zero hits, for exam­ ple, need not necessarily construe failure when a patron performs a quality search and finds that the library simply does not own the title in question. concerning intelligible outputs from transaction­log study, peters found that, “if the user is seen as carrying on a dialog of sorts with the online catalog, then it could be said that most transaction logs record only half of the conversa­ tion. more information about the system’s response to the user’s queries would help us better understand why patrons do what they do.” a look at subsequent transaction­log analyses into the s reveals somewhat differing research approaches yet strikingly similar results. wallace ( ) duplicated peters’s methods at eleven terminals within the university of colorado library system. her efforts spanned twenty hours of search monitoring and resulted in , logged searches. these were defined by carl system search type, (e.g., word, subject), then analyzed as cumulative totals and percentages of all searches. in this case, how­ michael j. bennett michael j. bennett (mbennett@cwmars.org) is digital initiatives librarian, c/w mars library network, worcester, massachusetts. article title | author opac design enhancements | bennett ever, failed searches (peters’s zero hits) were eliminated entirely from the sample as wallace focused primarily on patterns of completed searches and did not concern her­ self with questions of search success or failure, thus limit­ ing the scope of her findings. among searches analyzed, results were comparable to peters’s. in keeping with peters’s line of thinking, wallace remarked, intriguing vagaries in human behavior during an infor­ mation search process continue to stymie researchers’ efforts to understand that process. . . . current, widely used and described guidelines, rules and principles of searching simply do not take into account important aspects of what is really going on when an individual is using a computer to search for information. in , ciliberti et al. conducted a materials avail­ ability study of opac searches at adelphi university over a three­week period during fall semester. their work combined kantor’s branching­analysis methodol­ ogy with transaction­log analysis of opac use in order to better understand if users obtain the materials they need through the online catalog. sampling was accom­ plished during random open hours and drew informa­ tion from undergraduate, graduate, and faculty users. survey forms included questions of what patrons were searching for. forms were then picked randomly by staff for re­creation. the study was unclear as to the actual design of these forms and their queries. as a result their effectiveness remains questionable. a seven­category scheme was developed to code search failures that closely followed kantor’s branching analysis, where the concept of errors extends beyond just opac and its design to include such things as library collection devel­ opment and circulation practices. the survey itself along with the loss of accuracy that can be expected from patrons attempting to describe their searches on paper, then having these same searches re­created by research staff lead this author to question the data’s validity. as peters has noted, surveys are good for assessing opac users’ opinions but not necessarily their behavior. it would seem that in this instance the tool did not fit the task. this study did, however, use transaction logs after the initial survey analysis and indeed found discrepancies between the self­report (survey) and actual transaction­log data. search errors were subsequently categorized as pre­ viously described. though branching analysis is adept at examining on a holistic, entire­library scale (e.g., the ques­ tion of why patrons are not able to obtain materials), the method’s inherent breadth of focus does not lend itself to fine scrutiny of opac design issues in and of themselves. further refinement of the transaction­log analysis methodology may be seen in blecic’s et al. four­year longi­ tudinal study of opac use within the university of illinois library system. once again, failed searches, termed “zero postings” by the authors, were examined as dependent variables and percentages of the total number of searches and were used as a control. reasons for zero postings (e.g., searches missing search statements, author names entered in incorrect order) fell into seven separate catego­ ries. subsequent transaction­log sets were then culled after three incremental opac enhancements. enhancements included redesigns of general introductory and explain screens. z­test analysis of the level of equality between percentages of zero postings from log set to log set was then made in order to assess whether or not the enhance­ ments had any affect on diminishing said percentages and thus improving searching behavior. what blecic et al. found was temporary improve­ ment in patron searches followed by an unexpected lowering of patron performance over time. confounding attributes to the study include its longitudinal nature in an academic setting where user groups are not constant but variable. sadly, no attempt at tracking such possible changes in user populations was made. also of note was the fact that, as time passed, the command­based opac was increasingly being surrounded by web­based journal database search interfaces that did not require the use of sophisticated search statements and arguments. as users became accustomed to this type of searching, their com­ mand syntax skills may have suffered as a result. merits of the study include its straightforward design, logical data analysis, and plausible conclusions. longitudinal studies, though prone to the confound­ ing variables described, nevertheless form a persuasive template for further research into how incremental opac enhancements affect actual opac use over time. variations of transaction­log analysis also include the purely experimental. thomas’s simulation study of eighty­two first­year undergraduates at the university of pittsburg utilized four separate experimental screen inter­ faces. these interfaces included one that mimicked the current catalog with data labels and brief bibliographic displays, a second interface with the same bibliographic display but no data labels, and a third that contained the data labels but modified the brief display to include more subject­oriented fields. a fourth interface viewed the same brief displays as the third group but with the labels removed. users were pretested for basic demographic informa­ tion and randomly assigned to one of the four experi­ mental interface groups. each group was then given the same two search tasks. for the first task, users were asked to select items that they would examine further for a hypothetical research paper on big­band music and the music of duke ellington. the second task involved asking participants to examine twenty bibliographic records and to decide whether they would choose to look into these records further. participants were then asked to identify the data elements used to inform their information technology and libraries | march information technology and libraries | march relevance choices. resulting user behavior was subse­ quently tracked through transaction logs. for thomas’s experimental purposes, though, trans­ action logs took on a higher level of sophistication than in earlier comparative studies. here participants’ actions were monitored with a greater level of granularity. quantitative data were tracked for screens visited, time spent viewing them, total number of screens, total number of bibliographic citations examined at each level of speci­ ficity, and total time it took to complete the task. because of the obtrusive nature of the project, a third party was hired to administer the experiment. chi­square analysis of demographic data found no significance among partici­ pant groups in terms of their experience in using comput­ ers, online catalogs, or prior knowledge of the problem topic. this important analysis allowed the researchers a higher level of confidence in their subsequent findings. results in many instances were, however, inconclu­ sive. factors impairing the clarity of conclusions included the number of variables analyzed and the artificiality of the test design itself. thomas comments on one particular example of this: one of the fields that previous researchers said that library users found important was the call number field. obviously, without the call number, locating the actual item on the shelf is greatly complicated. in this experi­ ment, however, participants were not asked to retrieve the items they selected; thus, their perceived need for the call number may well have been mitigated. here is further evidence that a study of opac activity viewed in the context of actual outcomes, namely circula­ tion, is a logical approach to consider. most recently, graham at the university of lethbridge, alberta, examined opac subject searching and no­ hit results and considered two possible experimental enhancement types in order to allow users the ability to conduct more accurate searches. over a one­week period, , no­hit subject searches were first sampled and placed into nine categories by error type. subtotals were then expressed as percentage distributions of the total. a similar examination of , no­hit findings was also made over the course of four calendar years, form­ ing a longitudinal approach. percent distribution of error types from the two studies were then compared and were found to be similar with “non­library of congress subject headings” being the predominant area of concern. graham then attempted to improve subject searching by systematically enhancing the catalog in two ways. first, cross­references were created based upon the original no­ hit search term and linked to existing library of congress subject headings (lcshs) that graham interpreted as appropriate to the searcher’s original intentions. second, in instances where the original search could not be easily linked to an existing lcsh, a pathfinder record was cre­ ated that suggested alternate search strategies. all total, , new authority records and , pathfinder records were created over the course of the longitudinal study. the experiment, unfortunately, only went this far. no attempt was subsequently made to test whether these two methods of adding value to an existing opac search interface made a difference in users’ experiences. though creative in its suggested ameliorations to no­hit searches, the study also lacked any statistical testing of comparative data among sample years. possible problematic design issues, such as the relative complexity of pathfinders and how this might affect their end use were discussed but never tested through the analysis of real outcomes. in summary, major weaknesses of the transaction­log analysis model as demonstrated through the literature include: . lack of standardization among general study methodologies. . lack of standardization of opacs themselves: command structure and screen layout differ among software vendors. . lack of standards on measurable levels of search “success” or “failure.” while the following study of opac design enhance­ ments in the public library consortium environment did not directly address the first two points of emphasis, it was this author’s expectation that the lack of stan­ dardized notions of opac search success or failure found throughout the literature may be better addressed through a longitudinal analysis of discrete circulation and ill statistics. in this way, these quantifiable outcomes, both the direct results of patron initiation, would better assume clearer measures of patron success or failure in opac end use. ■ purpose and methodology in recent years, both academic and public libraries have invested substantial capital in improving opac design and automated systems. to what extent have these improvements affected the use of library materials by public library patrons? in order to better examine the question, this study tracked, over a seven­year period dating back from july through june , the circulation and systemwide holds statistical trends of sixteen member libraries of c/ w mars, a massachusetts automated library network of libraries. during this time a number of discrete, incre­ mental opac modifications granted patrons the ability to accomplish tasks remotely through the opac that previ­ ously had required library staff mediation. among these article title | author �opac design enhancements | bennett � changes, the initiation of intra­consortium (c/w mars) patron­placed holds, and the subsequent introduction of a link from the existing opac to the massachusetts virtual catalog (nine massachusetts consortiums, four university of massachusetts system libraries) were examined. this author hypothesized that such opac enhance­ ments that allow for broader choices of patron­placed holds would result in increases in both total circulation and total network transfers (ill) of library materials one year after initial enhancement adoption. as both total cir­ culation and total ill grew, it was hypothesized that ill as a percent of total circulation would likewise increase due to the fact that each opac enhancement was targeted directly toward facets of ill procurement. opac enhancements followed the schedule below: . general c/w mars network systemwide holds (requests mediated through library staff only), november . patron­placed holds (request button placed on c/ w mars opac screens), december . c/w mars participation in the massachusetts virtual catalog (additional button for pass through opac searches and requests from c/w mars catalog into the massachusetts virtual catalog), august these dates served as independent variables in a study of separate dependent variables (total circulation and total ills received) for all eight libraries one year after initial adoption of a new enhancement. for the sake of continu­ ity the terms holds and ills were used interchangeably throughout this examination. t­test comparisons to fig­ ures from the year prior to enhancement were then made for statistical significance. in addition, ills received as a percentage of total circulation (dependent variable) for all fifteen libraries one year after initial adoption of a new enhancement were also calculated and compared to the year prior to enhancement through z­test analysis. libraries chosen were a random sample from both central and western geographic regions of the network. sampled institutions did not go through any substantial renovations, drastic open hours changes, or closures dur­ ing the study period in order to better avoid potential con­ founding variables that may have skewed the resulting data. raw circulation and ill figures were taken directly from the massachusetts board of library commissioners’ (mblc) data files for fiscal years through . in the mblc’s data files, the following fields, sorted by library, correlated to this study’s statistical reporting: “dircirc” = “circulation” “loan from” = “ill” as fiscal year (fy) figures for circulation and ill had not yet been compiled by mblc at the time of this writing, these statistics were in turn taken directly from reports run off of c/w mars’s network servers. it should be noted that similar c/w mars reports are distributed and used by the consortium’s libraries them­ selves each fiscal year for reporting circulation and ill statistics to mblc. raw data by library were entered into microsoft excel spreadsheets. totals for circulation and ills received for all libraries by fy of opac enhancement were totaled and then compared to fy data prior to enhancement as a percent change value. excel’s data analysis tools were then employed to run t­tests (paired two sample for means) in tables through to analyze the level of change for significance from one sample to the next in both total circulation and total ills. (all tables and charts can be found in appendix following article.) tests for sig­ nificance employed two­tailed t­tests with an alpha level set to . . raw data for these same libraries across identical study years were also entered into subsequent spread­ sheets (tables through ) for additional z­tests (two samples for means) to analyze the level of change for significance from one fy sample to the next in ills received as a percentage of total circulation. here tests for significance employed two­tailed z­tests with an alpha level set to . . ■ results and discussion the results of a sixteen­library, seven­year longitudinal study of total circulation and total ills­received statistics are outlined in tables through , charts through . in addition, an analysis of ills received as a percentage of total circulation during this same time period among sampled libraries is represented in tables through . over the course of the study a total of , , circula­ tion and , ill transactions were examined from july through june . yearly comparisons in total circulation and total ills received from fy ’ to fy ’ were made to analyze the level of changes in circulation and ill statistics between years before any opac ill enhancements were under­ taken. as such these numbers gave insight into what changes, if any, normally occur in circulation and ill fig­ ures prior to a schedule of substantial opac ill enhance­ ments. although the year­to­year comparisons over the course of subsequent enhancement rollouts were made to test for the statistical significance of the year prior and following a particular functionality addition, the ’ to ’ information technology and libraries | march information technology and libraries | march comparison was made to form a control of what circula­ tion and ill trends may look like between years of no drastic workflow or design changes. results showed that this yearly comparison prior to the beginning of opac enhancements (table , charts and ) showed no significant change from one year to the next in total circulation (t = . , p > . ) or total ills received (t = ­ . , p > . ). circulation from ’ to ’ declined slightly by . percent while total ills received increased . percent. the mblc’s available retrospec­ tive data set currently only goes back to fy ’ , so a deeper understanding beyond this two­year comparison of normal year­to­year trends was impossible to achieve. yet data from this sample suggest that both circulation and ills may trend statistically flat from one year of little if any alteration of ill design to the next. additionally, comparisons of the percent of total ills received to total circulation were made between ’ and ’ (as will be seen in table ) and were found to be insignificantly different (z = ­ . , p > . ). ills received made up . percent of total circulation in fy ’ and . percent of total circulation in fy ’ . during fy ’ (november ), c/w mars rolled out automated systemwide holds functionality whereby library staff were first able to place patron requests for materials at other c/w mars member libraries through the consortium’s automated circulation system. up until this point, holds (ills) were placed primarily by staff through e­mail or faxed requests from one ill depart­ ment to another. patrons would request material either verbally with staff or through the submission of a paper or electronic form. staff would then look up the item in the electronic catalog and make the request. with the advent of systemwide holds, staff still accepted requests in a similar fashion, but instead of using the fax or e­mail, they began to place requests directly into the network’s innovative millennium circu­ lation clients. from there, the automated system not only randomly chose the lending library within the system but also automatically queued paging slips at the lending library for material that would subsequently be sent in transit to the borrowing location. by this time in the network’s development, opac had also graduated from a character­based telnet system to a smoother web design. but the catalog, in terms of directly assisting in the placing of ill requests, func­ tioned as it always had—it was still individually a search­ ing mechanism. the introduction of systemwide holds led to the sec­ ond largest jump in ill figures out of all comparative samples (table , chart ). interestingly enough, the con­ siderably significant . ­percent gain in ill activity from fy ’ to fy ’ (t = ­ . , p < . ) did not translate into a significant increase in total circulation. in fact, cir­ culation declined during this period, not significantly (t = . , p > . ), but by . percent nonetheless (table , chart ). a comparison of the percent of ills to total circulation from fy ’ to fy ’ (table ) indicated a sig­ nificant increase of . percent to . percent (z = ­ . , p < . ). more on the possible effects to circulation that rising levels of ills may elicit will be touched upon. though no statistical evaluations were made between fy ’ and fy ’ (as no novel ill changes were made over this period), it should be noted that during fy ’ the network first allowed patrons the ability, through opac, to log into their own accounts remotely. patrons were given the ability to set up a personal identification number and view such things as a list of their checked­ out items. patrons were also allowed to place checks next to such items and to renew these items remotely. fy ’ saw the original direct ill enhancement to opac. during this year patrons were first given the opportunity to directly place ill requests of their own (patron­placed holds) for material found in the catalog through the addition of an opac screen request button. up until this time, all material requests had been medi­ ated by library staff. comparative total circulation results from the year before enhancement to fy ’ (table , chart ) showed only a slightly significant . percent increase (t = ­ . , p < . ). ills­received figures (table , chart ), however, jumped by a considerable . percent margin (t = ­ . , p < . ), strongly suggesting that the opac request­ button addition and its facilitation of patron­placed holds had a positive effect upon total ill activity as was hypothesized. finally, total ills received as a percentage of total circulation increased slightly from fy ’ ( . percent) to fy ’ ( . percent) (table ) but did not rep­ resent a significant shift (z = ­ . , p > . ). the last augmentation to the network’s opac design that this study examined was an additional link for ills through the massachusetts virtual catalog. the massachusetts virtual catalog at the time of this study was an online union catalog of nine massachusetts net­ work consortia and four university of massachusetts system libraries. unlike the previous request­button enhancement that allowed for seamless patron­placed holds within the c/ w mars catalog, the massachusetts virtual catalog link was not a button but a descriptive hyperlink (can’t find the title you want here? try the massachusetts virtual catalog next!) from the network’s opac to the virtual catalog’s own dedicated opac interface. once there, patrons were required to login to the virtual catalog and re­create their search queries from scratch as previous searches were not automatically passed through to the second catalog. in essence, the virtual catalog acted as an additional step for patrons to take beyond c/w mars’s list of holdings to broaden their search for materials that the network’s member libraries did not own. article title | author opac design enhancements | bennett comparative figures for total circulation between fy ’ and fy ’ (table , chart ) when the virtual catalog link was added to the c/w mars opac screen found circulation down an insignificant . percent (t = . , p > . ), which ran counter to hypothesized expectations. total ills received between fy ’ and fy ’ (table , chart ), however, rose . percent, which proved to be a highly significant increase (t = ­ . , p < . ). additionally ills as a percent of total circulation rose from . percent in fy ’ to . percent in fy ’ (table ), which was sta­ tistically significant (z = ­ . , p < . ) and pointed to not only gains in ill itself after the introduction of the virtual catalog link but also to the ever increasing proportion of total circulation that ill activity accounted for. the final statistical comparison accomplished in this study was a look at what possible cumulative effect, if any, both opac enhancements may have had from the year before the first enhancement’s rollout (patron­placed holds request button) to one year after the latest addition (virtual catalog hyperlink from opac). in turn, com­ parative numbers for circulation and ills between fy ’ and fy ’ were examined. total circulation over this time (table , chart ) increased insignificantly by . percent (t = ­ . , p > . ). total ills received (table , chart ), how­ ever, increased by . percent, the highest significant increase of any two comparative samples (t = ­ . , p < . ). ills as a percent of total circulation also increased significantly from . percent in fy ’ to . percent in fy ’ (z = ­ . , p < . ) (table ). if one steps back and examines the various compari­ sons discussed up to this point, certain trends become evident. over the course of the seven­year study, total circulation remained relatively flat, oscillating slightly back and forth, year to year with only one significant increase that occurred after the introduction of patron­ placed holds in fy ’ . these results, excluding fy ’ , ran against hypothesized expectations that predicted that as ill enhancements were rolled out, correspondingly significant increases in circulation would result. total ills received (the fy ’ to fy ’ control com­ parison) before the advent of first, network systemwide holds, then a succession of opac design enhancements that allowed for a broader range of patron­initiated ills suggested that these totals run statistically flat from one year to the next. with the advent of systemwide holds, the ill picture, however, began to change dramatically with a significant increase in total ills. this was fol­ lowed by significant increases in ill activity in each study year that came after an opac ill enhancement. these results pointed toward the substantial effect that these enhancements made in total ill activity and sup­ ported hypothesized expectations. when such opac rollouts were examined as a cumu­ lative influence through the prism of ill levels of this past fiscal year (fy ’ ) compared to the year before their initial advent (fy ’ ), the positive effect that such enrich­ ments had on not only total ill but also on total circula­ tion becomes clearest. for it is through this comparison that it was found that not only did total ills increase significantly but that ills as a percentage of total circula­ tion also increased significantly from the time before the first opac enhancement to the present. total circulation was surprisingly impervious to change and ran statisti­ cally flat during this time. it is clear from this longitudinal study that incremen­ tally granting patrons access to online tools for them to initiate such traditional library business as ills spurs sig­ nificantly large increases in such activity. in other words, these online tools are not ignored but are intellectually and literally grasped. what may be surprising, however, is the degree to which ill has increased as a result of them, to a point where ill has not only taken up a sig­ nificantly greater proportion of total circulation than ever before but also appears to be changing the very nature of circulation itself. future studies may include a deeper examination of the circulation and ill statistical picture farther back in time than this investigation covers to better clarify trends leading up to such major enhancement rollouts. also, similar longitudinal studies from different consortia envi­ ronments may shed further light on evidence discussed throughout this writing. consortia are uniquely poised to offer large statistical sample sizes and standardized workflows within their network­wide ill and circulation software packages and automated statistical programs. this, in turn, results in high­quality, consistent data samples from heterogeneous library sources that are rela­ tively uncorrupted by scattershot recording methods and differing circulation and ill methodologies. finally, a future look at the effects that similar opac ill enhancements may have on borrowing trends beyond general raw transactional figures is warranted. chris anderson, for example, has recently commented on long tail statistical analysis and its relation to library catalogs. here outwardly shifting demand curves for library mate­ rials are hypothesized as collections become more visible and interconnected through the web. in a similar vein, a more granular examination of such concepts as possible circulation and ill­activity trends in terms of discrete material types borrowed, patron types who borrow, or a cross­tabulation of these data points would appear to be a fertile next step toward a greater knowledge of ills and circulation as a whole. references . t. peters, “when smart people fail: an analysis of the transaction log of an online public access catalog,” the journal of academic librarianship , no. ( ): – . information technology and libraries | march information technology and libraries | march . ibid., . . ibid. . ibid., . . p. wallace, “how do patrons search the online catalog when no one’s looking? transaction­log analysis and impli­ cations for bibliographic instruction and system design,” rq , no. ( ): – . . peters, “when smart people fail.” . wallace, “how do patrons search the online catalog when no one’s looking?” . . a. ciliberti et al., “empty handed? a material availabil­ ity study and transaction­log analysis verification,” the journal of academic librarianship , no. ( ): – . . p. kantor, “availability analysis,” journal of the american society for information science , nos. – ( ): – . . ciliberti et al., “empty handed? a material availability study and transaction­log analysis verification.” . peters, “when smart people fail.” . ciliberti et al., “empty handed? a material availability study and transaction­log analysis verification.” . d. blecic, et al., “a longitudinal study of the effects of opac screen changes on searching behavior and searcher suc­ cess,” college & research libraries , no. ( ): – . . ibid. . d. thomas, “the effect of interface design on item selec­ tion in an online catalog,” library resources & technical services , no. ( ): – . . ibid., . . r. graham, “subject no­hits searches in an academic library online catalog: an exploration of two potential ame­ liorations,” college & research libraries , no. ( ): – . . ibid. . massachusetts board of library commissioners , “public library data, data files,” http://www.mlin.lib.ma.us/ advisory/statistics/public/index.php (accessed oct. , ). . c. anderson, “the long tail,” wired magazine , no. ( ): – ; “q&a with chris anderson,” oclc newsletter, , no. , http://www.oclc.org/news/publications/news letters/oclc/ / /interview.htm (accessed july , ). appendix a: tables and charts table . yearly comparison prior to the beginning of ill opac enhancements table . general systemwide holds implementation (adopted / ) article title | author opac design enhancements | bennett table . opac design enhancement: patron-placed holds (adopted / ) table . opac design enhancement: patron-placed massachusetts virtual catalog holds (adopted / ) table . opac design enhancements: “cumulative effect” (fy ’ to fy ’ ) table . yearly comparison prior to the beginning of ill opac enhancements of ill received as a percentage of total circulation information technology and libraries | march information technology and libraries | march table �. opac design enhancement: patron-placed massachusetts virtual catalog holds (adopted / ) ill received as a percentage of total circulation table . opac design enhancements: “cumulative effect” (fy ’ to fy ’ ) ill received as a percentage of total circulation table . general systemwide holds (adopted / ) ill received as a percentage of total circulation table . opac design enhancement: patron-placed holds (adopted / ) ill received as a percentage of total circulation article title | author opac design enhancements | bennett chart . circulation comparison prior to any ill opac enhance- ment (fy ’ to fy ’ ) chart . ill received comparison prior to any ill opac enhance- ment (fy ’ to fy ’ chart . holds received comparison before and after general systemwide holds implementation (adopted / ) chart . circulation comparison before and after patron-placed holds opac enhancement (adopted / ) chart . circulation comparison before and after general systemwide holds implementation (adopted / ) chart . holds received comparison before and after patron-placed holds opac enhancement (adopted / ) information technology and libraries | march information technology and libraries | march chart . circulation comparison before and after massachusetts virtual catalog opac enhancement (adopted / ) chart . holds received comparison before and after massachusetts virtual catalog opac enhancement (adopted / ) chart . circulation comparison opac enhancements “cumulative effect” (fy ’ to fy ’ ) chart . ill comparison opac enhancements “cumulative effect” (fy ’ to fy ’ ) lita , , cover , cover neal­schuman cover index to advertisers information technology and libraries | march author id box for column layout column title editor zoomify image is a mature product for easily publishing large, high-reso- lution images on the web. end users view these images with existing web- browser software as quickly as they do normal, downsampled images. a flash-based zoomifyer client asyn- chronously streams image data to the web browser as needed, resulting in response times approaching those of desktop applications using minimal bandwidth. the author, a librarian at cornell university and the principal architect of a small, open-source com- pany, worked closely with zoomify to produce a cross-platform, open- source implementation of that com- pany’s image-processing software and discusses how to easily deploy the product into a widely used web- publishing environment. limitations are also discussed as are areas of improvement and alternatives. z oomifyer from zoomify (www .zoomify.com) enables users to view large, high­resolu­ tion images within existing web­ browser software while providing a rich, interactive user experience. a small zoomifyer client, authored in macromedia flash, is embedded in an html page and makes asyn­ chronous requests to the server to stream image data back to the client as needed. by streaming the image data in this way, the image renders as quickly as a normal, downsampled image, even for images that are giga­ bytes in size. as the user pans and zooms, the response time approaches that of desktop applications while using the smallest possible band­ width necessary to render the image. and because flash has . per­ cent browser saturation, viewing “zoomified” images is seamless for most users and allows them to view images interactively in much greater detail than would otherwise be prac­ tical or even possible. zoomify image (sourceforge.net/ projects/zoomifyimage) was created at cornell university in collabora­ tion with zoomify to create an open­ source, cross­platform, and scriptable version of the processing software that creates the image data displayed in a zoomifyer client. this work was immediately integrated into an inno­ vative content­management system that was being developed within the zope application server, a premier web application and publishing plat­ form. authors in this system can add high­resolution images just as they normally add downsampled images, and the image is automat­ ically processed on the server by zoomify image and displayed within a zoomifyer client. zoomify image is now in its second major release on source forge and contains user con­ tributed software to easily deploy it in other environments such as php. zoomifyer has been used in a number of applications in many fields, and can greatly enhance many research and instructional activities. applying zoomifyer to digital­image collections is obvious, allowing libraries to deliver an unprecedented level of detail in images published to the web. new applications also suggest themselves, such as serving high­resolution images taken from tissue samples in a medical lab or using zoomifyer in advanced geo­ spatial image applications, particu­ larly when advanced client features such as annotations are used. the zoomifyer approach also has positive implications for preservation and copyright protection. zoomify image generates cached derivatives of master image files so the image masters are never directly accessed in the application or sent over the internet. image data are stored and transmitted to the client in small chunks so that end users do not have access to the full data of the original image. deploying zoomify image dependencies and winstal- lation zoomify image was designed ini­ tially to be a faithful, cross­platform port of zoomify’s image­processing software. it was developed in close cooperation with zoomify to pro­ vide a scriptable method for invok­ ing the image­preparation process for zoomifyer clients so this technol­ ogy could be used in more environ­ ments. zoomify image is written in the python programming language and uses the third­party python imaging library (pil) with jpeg support, both of which are also open source and cross­platform. it has been tested in the following environments: ■ python . . ■ pil . . and ■ python . . ■ pil . . installers for python and pil exist for all major platforms and can be obtained at python.org and www .pythonware.com/products/pil. the installation documentation that comes with pil will help you locate the appropriate jpeg libraries if they are missing from your system. for macosx, you can find pre­built binary installers for python, pil and zope at sourceforge.net/projects/ mosxzope. introducing zoomify image adam smith adam smith (ajs @cornell.edu) is a systems librarian at cornell university library, ithaca, new york. introducing zoomify image | smith �introducing zoomify image | smith � the “ez” version of the zoomifyer client, a flash­based applet with basic pan and zoom functionality, is pack­ aged with zoomify image for conve­ nience so the software can be used immediately once installed. the ez client is covered by a separate license and can be easily replaced with more advanced clients from zoomify at www.zoomify.com. (a description of how to upgrade the zoomifyer client is included in this paper.) after python and pil with jpeg support are installed, download the zoomify image software from sourceforge.net/projects/zoomify­ image and decompress it. using zoomify image from the command line begin exploring zoomify image by invoking it on the command line: python /zoomifyfilepr ocessor.py or, to process more than one file at a time: python /zoomifyfile processor.py the file format of the images input to zoomify image are typically either tiff or jpeg, but can be any of the many formats that pil can read. an image called “test.jpg” is included in the zoomify image distribution and is of sufficient size and complexity to provide an interesting example. during processing, zoomify image creates a new directory to hold the converted image data in the same location as the image file being processed. the name of this direc­ tory is based on the file name of the image being processed, so that, for example, an image called “test.jpg” would have a corresponding folder called “test” containing the converted image data used by the zoomifyer client. if the image file has no file extension, the directory is named by appending “_data” to the image name, so that an image file named “test” would have a corresponding directory called “test_data.” if the process is re­run on the same images, any previously generated data are automatically deleted before being regenerated. zoomify provides substantial documentation and sample code on its web site that demonstrates how to use the data generated by zoomify image in several environments. user­ contributed code is bundled with zoomify image itself, further dem­ onstrating how to dynamically incor­ porate this conversion process into several environments. an example of the use of zoomify image within the zope application server is given. incorporating zoomify image into the zope application server the popular zope application server contains a number of built­in services including a web server, ftp and webdav servers, plug­ins for access­ ing relational databases, and a hier­ archical object­oriented database that uses a file­system metaphor for stor­ age. this object database provides a unique opportunity to incorporate zoomifyer into zope seamlessly. to use zoomify image with zope, the distribution must be decom­ pressed into your zope products directory. for versions . .x and up, this is at: /products/ in zope versions prior to the . .x series, the products directory is at: /lib/python/ products/ restart zope and now within the web­based zope management interface (zmi), the ability to add zoomify image objects appears. after selecting this option, a form is presented that is identical to the form used for adding ordinary image objects within zope. when an image is uploaded using this form, zope automatically invokes the zoomify image conversion process on the server and links the generated data to the default zoomifyer client that comes with the distribution. if the image is subsequently edited within zmi to upload a new version, any existing conversion data for that image are automatically deleted, and the new conversion data are gener­ ated to replace them, just as when invoked on the command line. again, the uploaded image can be in any format that zope recognizes as having a content­type of “image/...” and that pil can read. the only potential “gotcha” in this process is that in the versions of the zoomifyer client the author has tested, zoomify image objects that have file names (in zope terminology, the file name is the object’s “id” property) with extensions other than “.jpg” are not displayed properly by the zoomifyer client. so, when uploading a tiff image, for example, the id given to the zoomify image object should either not contain an extension, or it should be changed from image.tif to something like image_tif. this bug has been reported to zoomify and may be fixed in newer versions of the flash­based viewing software at the time of publication. to view the image within the zoomifyer client, simply call the “view” method of the object from within a browser. so, for a zoomify image object uploaded to: http:///test/test.jpg go to this url: http:///test/test. jpg/view or, to include this view of the image within a zope page template information technology and libraries | march information technology and libraries | march (zpt), simply call the tag method of the zoomify image just as you would a normal image object in zope. so, in a zpt, use this: it is possible that the zoomify image conversion process will not have had time to complete when someone tries to view the image. the zoomify image object will attempt to degrade gracefully in this situation by trying to display a downsampled version of the image that is gener­ ated part way through the conver­ sion process, or, if that is also not available, finally informing the user that the image is not yet ready to be viewed. this logic is built into the tag method. to add larger images more effi­ ciently, or to add images in bulk, the zoomify image distribution contains detailed documentation to quickly configure zope to accept images via ftp or webdav and automatically process them through zoomify image when they are uploaded. finally, the default zoomifyer cli­ ent can be overridden by uploading a custom zoomifyer client into a loca­ tion where the zoomify image object can “acquire” it, and giving it a zope id of “zoomifyclient.swf”. how it works to be viewed by a zoomifyer cli­ ent, an image must be processed to produce tiles of the image at differ­ ent scales, or tiers. an xml file that describes these tiles is also necessary. zoomify image provides a cross­ platform method of producing these tiled images and the xml file that describes them. beginning at ­percent scale, the image is successively scaled in half to produce each tier, until both the width and height of the final tier are, at most, pixels each. each tier is further divided into tiles that are, at most, pixels wide by pixels tall, as seen in figure . these tiles are created left to right, top to bottom. tiles are saved as images with the naming conven­ tion indicated in figure . the numbering is zero­based, so that the smallest tier is represented by one tile that is at most x pixels wide with the name “ ­ ­ .jpg.” tiles are saved in directories in groups of , and those directories also follow a zero­based naming con­ vention starting with “tilegroup .” lower­numbered tile groups contain lower­numbered tiles, so ­ ­ .jpg is always in tilegroup . zoomifyer clients understand this tile­naming scheme and only request tiles from the server that are necessary to stitch together the por­ tion of the image being viewed at a particular scale. limitations zoomify image was developed to meet two goals: . to provide a cross­platform port of the zoomifyer con­ figure . tiers and tiles for a x pixel image figure . tile image naming scheme introducing zoomify image | smith introducing zoomify image | smith verter for use in unix/linux systems, and . to make the converter script­ able, and ultimately integrate it into open­source content­man­ agement software, particularly zope. this zoomifyer port was writ­ ten in python, a mature, high­level programming language with an execution model similar to java. although zoomify image continues to be optimized, compared to the official zoomify conversion software, it is slower and more limited in the sizes of images it can reasonably process. anecdotally, zoomify image has been used effectively on images hundreds of megabytes large, but significant performance degradation has been reported in the multi­giga­ byte range. because of these limitations in zoomify image, the official zoomify image­processing software is recom­ mended for converting very large images manually in a windows or macintosh environment. the zoomify image product is recommended in the following circumstances: ■ the conversion must be per­ formed on a unix/linux machine. ■ the conversion process must be scriptable, such as for batch pro­ cessing or being run dynamically. ■ images sizes are not in the multi­ gigabyte range. if a scriptable, cross­platform version of the zoomifyer converter is needed, but performance is an issue, several things can be done to extend the current limits of the soft­ ware. obviously, upgrading hard­ ware, particularly ram, is effective and relatively inexpensive. running the latest versions of python and pil will also help. each new version of python makes significant perfor­ mance improvements, and this was a primary goal of version . , which was released in september . the author believes that the cur­ rent weak link in the performance chain is related to how zoomify image is loading image data into memory with pil during processing. in the current distribution, a python script contributed by gawain avers, which is based partially on the zoomify image approach, uses imagemagick instead of pil for image manipula­ tion and is better able to process multi­gigabyte images. the author would like to add the ability to des­ ignate the image library at runtime in future versions of zoomify image. future development beyond improving the performance of the core­processing algorithm, the author would also like to explore opportunities for more efficiently processing images within zope, such as spawning a background thread for processing images so the zope web server can immediately respond to the client’s image­submission request. the author would also like to improve the tag method to display data more flexibly in the zoomifyer client and ensure consistent behav­ ior with zope’s default image tag method. finally, zoomify image could also benefit from the addi­ tion of a simple configuration file to control such runtime properties as image quality and which third­party image­processing library to use, for example. conclusion zoomify image is mature, open­ source software that makes it pos­ sible to publish large, high­resolution images to the web. it is designed to be convenient to use in a variety of architectures and can be viewed within existing browser software. download it for free, begin using it in minutes, and explore its unique possibilities. references . adobe systems, macromedia flash player statistics, http://www.adobe.com/ products/player_census/flashplayer/ (accessed march , ). . pythonware, python imaging library handbook: image file formats, http:// www.pythonware.com/library/pil/ handbook/formats.htm (accessed aug. , ). resources macromedia flash player statistics (http://www.adobe.com/ products/player_census/flash­ player/) (accessed jan. , ). python imaging library (pil) (http:// www.pythonware.com/products/ pil/) (accessed jan. , ). python programming language official web site (http://www.python.org/) (accessed jan. , ). zoomify image (http://sourceforge.net/ projects/zoomifyimage/) (accessed jan. , ). zoomify (http://www.zoomify.com/) (accessed jan. , ). zope community (http://www.zope .org/) (accessed jan. , ). zope installers for macosx (http:// sourceforge.net/projects/ mosxzope/) (accessed jan. , ). ital_ n p ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ reproduced with permission of the copyright owner. further reproduction prohibited without permission. wikiwikiwebs: new ways to communicate in a web environment chawner, brenda;lewis, paul h information technology and libraries; mar ; , ; proquest education journals pg. reproduced with permission of the copyright owner. further reproduction prohibited without permission. reproduced with permission of the copyright owner. further reproduction prohibited without permission. reproduced with permission of the copyright owner. further reproduction prohibited without permission. reproduced with permission of the copyright owner. further reproduction prohibited without permission. reproduced with permission of the copyright owner. further reproduction prohibited without permission. reproduced with permission of the copyright owner. further reproduction prohibited without permission. reproduced with permission of the copyright owner. further reproduction prohibited without permission. reproduced with permission of the copyright owner. further reproduction prohibited without permission. reproduced with permission of the copyright owner. further reproduction prohibited without permission. reproduced with permission of the copyright owner. further reproduction prohibited without permission. graphical table of contents for library collections: the application ... herrero-solana, victor;félix moya-anegón;guerrero-bote, vicente;zapico-alonso, felipe information technology and libraries; mar ; , ; proquest education journals pg. reproduced with permission of the copyright owner. further reproduction prohibited without permission. reproduced with permission of the copyright owner. further reproduction prohibited without permission. reproduced with permission of the copyright owner. further reproduction prohibited without permission. reproduced with permission of the copyright owner. further reproduction prohibited without permission. ital_ n p ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ital_ n p - ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ital_ n .pdf ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ital_ n p - ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ ฀ reproduced with permission of the copyright owner. further reproduction prohibited without permission. wikiwikiwebs: new ways to communicate in a web environment chawner, brenda;lewis, paul h information technology and libraries; mar ; , ; proquest education journals pg. reproduced with permission of the copyright owner. further reproduction prohibited without permission. reproduced with permission of the copyright owner. further reproduction prohibited without permission. reproduced with permission of the copyright owner. further reproduction prohibited without permission. reproduced with permission of the copyright owner. further reproduction prohibited without permission. reproduced with permission of the copyright owner. further reproduction prohibited without permission. reproduced with permission of the copyright owner. further reproduction prohibited without permission. reproduced with permission of the copyright owner. further reproduction prohibited without permission. reproduced with permission of the copyright owner. further reproduction prohibited without permission. reproduced with permission of the copyright owner. further reproduction prohibited without permission. reproduced with permission of the copyright owner. further reproduction prohibited without permission. reproduced with permission of the copyright owner. further reproduction prohibited without permission. graphical table of contents for library collections: the application ... herrero-solana, victor;félix moya-anegón;guerrero-bote, vicente;zapico-alonso, felipe information technology and libraries; mar ; , ; proquest education journals pg. reproduced with permission of the copyright owner. further reproduction prohibited without permission. reproduced with permission of the copyright owner. further reproduction prohibited without permission. reproduced with permission of the copyright owner. further reproduction prohibited without permission. reproduced with permission of the copyright owner. further reproduction prohibited without permission. information technology and libraries | june author name and second author f orty years! in july , the library and informa- tion technology association (lita) was officially born at the american library association (ala) annual conference in new york as the information science and automation division (isad). it was bastille day, and i’m sure for those who had worked so hard to create this new organization that it probably seemed like a revolution, a new day. the organizational meeting held that day attracted “several hundred people.” imagine! i’ve mentioned it before, i know, but the history of the first twenty-five years of lita is intriguing reading and well worth an investment of your time. stephen r. salmon’s article “lita’s first twenty-five years: a brief history” (www.lita.org/ala/lita/aboutlita/org/ st years.htm) offers an interesting look back in time. any technology organization that has been in existence for forty or more years has seen a lot of changes and adapted over time to a new environment and new technologies. there is no other choice. someone (who, i don’t remember; i’d gladly attribute the quote if i did) once told me that library automation began with the electric eraser. i’m sure that many of you have neither seen an electric eraser, nor can you probably imagine its purpose. ask around. i’m sure there are staff in your organization who do remember using it. there may even be one hidden somewhere in your library. a quick search of the web even finds cordless, rechargeable electric erasers today in drafting and art supply stores. the s, as lita was born, was still the era of the big mainframe systems and not-so-common program- ming languages. machine readable cataloging (marc) was born and oclc conceived. the s saw the intro- duction of minicomputer systems. digital equipment corporation introduced the vax, a -bit platform, in . the roots of many of our current integrated library systems reach back to this decade. the s saw the introduction of the ibm personal computer and the apple macintosh. the graphical interface became the norm or at least the one to imitate. the s saw a shift away from hardware to communication and access as the web was unveiled and began to give life to the internet bubble. the new millennium began with y k. the web predomi- nates, and increasingly, the digital form dominates almost everything we touch (text, audio, video). automation and systems evolved and changed over the years, and so did libraries. automation, which had been confined to large air-conditioned and moni- tored rooms, moved out into the library. it increas- ingly appeared at circulation desks, on staff desks, and then throughout the library. information technology (it) spread into offices everywhere and into homes. libraries had products and services to deliver to users. users demanded more convenience. of course, others knew this trend as well and provided products and services that users wanted. users often liked what they saw in stores better than what the library was able to provide. each of us attempts to keep up, compete, and beat those whom we see as our competitors. it’s a moving target and one that seems to be gaining speed. all the while, during these four decades, our asso- ciation and its members continually adapted to the new environment, faced new challenges, and adopted new technologies. we would not exist if we did not. i feel that we, as an association, are again facing the need to change, to trans- form ourselves. it, digital technology, automation (whatever term you want to use) affects the work of virtually every library staff member. everyone’s work in the library uses or con- tributes to the digital presence of our employer. it is not the domain of a few. lita has a wonderful history and it has great poten- tial to better serve the profession. what do we want our association to be? what programs and services can we provide that others do not? who can we involve to broaden our reach? how can we better communicate with members and nonmembers? if we had a clean sheet of paper, what would we write? what would we dream? we need to share that dream and bring it to life. i can’t do it. the lita board can’t do it. we need your help. we need your ideas. we need your energy. we need to break out of our comfort zone. none of us wants the strategic plan (www.lita.org/ala/lita/aboutlita/org/plan.htm) we adopted last year to ring hollow. we want to accelerate change and move into a reenergized future. i welcome your aspirations, ideas, and comments. i know that the lita board does as well. please feel free to contact me or any member of the board (www.lita .org/ala/lita/aboutlita/org/litagov/board.htm). lita is your association. where should we be going? help us navigate the future. patrick mullin patrick mullin (mullin@email.unc.edu) is lita president – , and associate university librarian for access services and systems, the university of north carolina at chapel hill. president’s column author name and second author b y now, most library and information technology association (lita) members and information tech- nology and libraries (ital) readers know that is the fortieth anniversary of lita’s predecessor, the information science and automation division (isad) of the american library association (ala). and marks the fortieth birthday of ital, first published in as the journal of library automation (jola). i hope that members and readers know the vital role played by fred kilgour in the founding of the division and as jola’s founding editor. this issue marks the initiation of a two-volume cel- ebration (volumes and ) of his role as founding edi- tor by publishing what we hope are significant articles resulting from original research, the development of important and creative new systems, or explications of significant new technologies that will shape future infor- mation technologies. i have invited some of the authors of these articles to submit their manuscripts. others are being submitted in response to a call i published both in an earlier editorial and in a message to the lita-l discus- sion list. whether invited or submitted, they will receive the same double-blind refereeing that all ital articles undergo. the referees will not know which articles have been invited or submitted for this purpose. the articles will, however, be so designated when they are published. volume initiates a second landmark for ital. henceforth, ital will be published simultaneously in electronic and print versions. the electronic copy will be available to lita members and ital subscribers on the ala/lita web site. equally significantly, at the ala midwinter meeting in san antonio, the lita board of directors approved a second proposal from the lita publications committee. (the ital editor and edito- rial board report to the publications committee.) after six months, the electronic issues will be open to all, not restricted to members and subscribers. put simply, if you are a member or subscriber read- ing this issue in print, you may also read it and volume , number (the march issue) on the web. when volume , number is published in september , the march issue on the web will be open for anyone to read. when the december issue is published, this june e-issue will be open to all. the web versions are to be published in both pdf and html versions. most ital articles now include urls. readers will be able to link to them. most figures and graphs submitted by authors are in color. from now on, these will be available to the readers of the e-copies. ala publishing allows authors to submit their arti- cles to institutional repositories, and many authors now do so. authors will retain this option. some articles have been posted on other portals as well. martha yee’s out- standing june article on how to frbrize the opac appears not only on ucla’s repository site but also on the escholarship repository site of the university of california system, one of the few library-related articles on the site (http://repositories.cdlib.org/escholarhip). furthermore, on november , , it was among the top ten most popular articles on the site. recently, dlist (http://dlist.sir.arizona.edu) at the university of arizona library received permission to include it. the decisions to allow simultaneous publication of print and electronic versions and to allow open access after six months were not made lightly. the lita board members carried on extensive electronic discussions among themselves and with nancy colyar, chair of the publications committee, and me. lita president pat mullin’s summary of those discussions was more than ten single-spaced pages. nancy and i also attended a meeting of the board in san antonio. publications and memberships are two chief sources of revenue for almost all professional associations. in two surveys in the past ten years, lita members have indicated they considered ital to be their most important membership benefit. lita membership fell this year, probably because of the recent dues increases by other divisions of ala. this decline was anticipated by lita’s leadership. i think both the ital editorial board and the lita leadership would love to take the additional pioneering step of making our journal a full open-access publication. however, legitimate concern was expressed that opening access after six months might lead to both a decrease in members and subscribers. a significant number of lita leaders said that their membership was based on lita programs, participation, and interaction with colleagues, not just ital. i hope that all lita members feel the same. i further hope that lita members will do everything they can to discourage their libraries from canceling their sub- scriptions. our financial health would be enhanced if all lita members took two other steps: participating in writ- ing and encouraging the writing of significant articles, and encouraging your many library technology vendors to advertise in ital. fred kilgour and the other founders of our division were library information technology (it) pioneers. fred’s leadership helped make jola and now ital vital read- ing for library it professionals. i believe that by celebrat- ing the lita/ital anniversaries with a reconfirmation of our practice of publishing articles of the highest quality and by making ital more accessible through electronic publication, we are reaffirming the scholarly and profes- sional commitments first made by fred kilgour and his isad colleagues such a short forty years ago. john webb john webb (jwebb@wsu.edu) is assistant director for systems and planning, washington state university libraries, pullman, and editor of information technology and libraries. editorial: lita and ital: forty and still counting editorial | webb author name and second author the use of ajax, or asynchronous javascript + xml, can result in web applications that demonstrate the flexibility, responsiveness, and usability traditionally found only in desktop software. to illustrate this, a repository metasearch user interface, ojax, has been developed. ojax is simple, unintimidating but power- ful. it attempts to minimize upfront user investment and provide immediate dynamic feedback, thus encouraging experimentation and enabling enactive learning. this article introduces the ajax approach to the develop- ment of interactive web applications and discusses its implications. it then describes the ojax user interface and illustrates how it can transform the user experience. w ith the introduction of the ajax development paradigm, the dynamism and richness of desk- top applications become feasible for web-based applications. ojax, a repository metasearch user inter- face, has been developed to illustrate the potential impact of ajax-empowered systems on the future of library software. this article describes the ajax method, highlights some uses of ajax technology, and discusses the implica- tions for web applications. it goes on to illustrate the user experience offered by the ojax interface. ■ ajax in february , the term ajax acquired an additional meaning: asynchronous javascript + xml. the con- cept behind this new meaning, however, has existed in various forms for several years. ajax is not a single technology but a general approach to the development of interactive web applications. as the name implies, it describes the use of javascript and xml to enable asyn- chronous communication between browser clients and server-side systems. as explained by garrett, the classic web application model involves user actions triggering a hypertext trans- fer protocol (http) request to a web server. the latter processes the request and returns an entire hypertext markup language (html) page. every time the client makes a request to the server, it must wait for a response, thus potentially delaying the user. this is particularly true for large data sets. but research demonstrates that response times of less than one second are required when moving between pages if unhindered navigation is to be facilitated through an information space. the aim of ajax is to avoid this wait. the user loads not only a web page, but also an ajax engine written in javascript. users interact with this engine in the same way that they would with an html page, except that instead of every action resulting in an http request for an entire new page, user actions generate javascript calls to the ajax engine. if the engine needs data from the server, it requests this asynchronously in the back- ground. thus, rather than requiring the whole page to be refreshed, the javascript can make rapid incre- mental updates to any element of the user interface via brief requests to the server. this means that the traditional page-based model used by web applications can be abandoned; hence, the pacing of user interaction with the client becomes independent of the interaction between client and server. xmlhttprequest is a collection of application pro- gramming interfaces (apis) that use http and javascript to enable transfer of data between web servers and web applications. initially developed by microsoft, xmlhttprequest has become a de facto standard for javascript data retrieval and is implemented in most modern browsers. it is commonly used in the ajax para- digm. the data accessed from the http server is usually in extensible markup language (xml) but another for- mat, such as javascript object notation, could be used. applications of ajax google is the most significant user of ajax technology to date. most of its recent innovations, including gmail, google suggest, google groups, and google maps, employ the paradigm. the use of ajax in google suggest improves the tradi- tional google interface by offering real-time suggestions as the user enters a term in the search field. for example, if the user enters xm, google suggest might offer refine- ments such as xm radio, xml, and xmods. experimental ajax-based auto-completion features are appearing in a range of software. shanahan has applied the same ideas to the amazon online bookshop. his experimental site, zuggest, extends the concept of auto-completion: as the user enters a term, the system automatically triggers a search without the need to hit a search button. the potential of ajax to improve the responsiveness and richness of library applications has not been lost on the library community. several interesting experiments have been tried. at oclc, for example, a “suggest-like service,” based on controlled headings from the world- judith wusteman and pádraig o’hiceadha using ajax to empower dynamic searching | wusteman using ajax to empower dynamic searching judith wusteman (judith.wusteman@ucd.ie) is a lecturer in the ucd school of information and library studies, university college dublin, ireland. information technology and libraries | june wide union catalog, worldcat, has been implemented. ajax has also been used in the oclc deweybrowser. the main page of this browser includes four iframes, or inline frames, three for the three levels of dewey decimal classification and a fourth for record display. the use of ajax allows information in each iframe to be updated independently without having to reload the entire page. implications of ajax there have been many attempts to enable asynchronous background transactions with a server. among alter- natives to ajax are flash, java applets, and the new breed of xml user-interface language formats such as xml user interface language (xul) and extensible application markup language (xaml). these all have their place, particularly languages such as xul. the latter is ideal for use in mozilla extensions, for example. combinations of the above can and are being used together; xul and ajax are both used in the firefox extension version of google suggest. the main advan- tage of ajax over these alternative approaches is that it is nonproprietary and is supported by any browser that supports javascript and xmlhttprequest—hence, by any modern browser. it could be validly argued that complex client-side javascript is not ideal. in addition to the errors to which complex scripting can be prone, there are accessibility issues. best practice requires that javascript interaction adds to the basic functionality of web-based content that must remain accessible and usable without the javascript. an alternative non-javascript interface to gmail was recently implemented to deal with just this issue. a move away from scripting would, in theory, be a positive step for the web. in practice, however, proce- dural approaches continue to be more popular; attempts to supplant them, as epitomized by xhtml . , simply alienate developers. it might be assumed that the use of ajax technol- ogy would result in a heavier network load due to an increase in the number of requests made to the server. this is a misconception in most cases. indeed, ajax can dramatically reduce the network load of web appli- cations, as it enables them to separate data from the graphical user interface (gui) used to display it. for example, each results page presented by a traditional search engine delivers, not only the results data, but also the html required to render the gui for that page. an ajax application could deliver the gui just once and, after that, deliver data only. this would also be pos- sible via the careful use of frames; the latter could be regarded as an ajax-style technology but without all of ajax’s advantages. ■ from client-server to soa the dominant model for building network applications is the client/server approach, in which client software is installed as a desktop application and data generally reside on a server, usually in a database. this can work well in a homogenous single-site computing environ- ment. but institutions and consortia are likely to be het- erogeneous and geographically distributed. pcs, macs, and cell phones will all need access to the applications, and linux may require support alongside windows. even if an organization standardizes solely on windows, different versions of the latter will have to be supported, as will multiple versions of those ubiquitous dynamic link libraries (dlls). indeed, the problems of obtaining and managing conflicting dlls have spawned the term “dll hell.” in web applications, a standard client, the browser, is installed on the desktop but most of the logic, as well as the data, reside on the server. of course, the browser developers still have to worry about “dll hell,” but this need not concern the rest of us. “speed must be the overriding design criterion” for web pages. but the interactivity and response times possible with client/server applications are still not avail- able to traditional web applications. this is where ajax comes in: it offers, to date, the best of the web application and client/server worlds. much of the activity is moved back to the desktop via client-side code. but the advan- tages of web applications are not lost: the browser is still the standard client. service-oriented architecture (soa) is an increas- ingly popular approach to the delivery of applications to heterogeneous computing environments and geo- graphically dispersed user populations. soa refers to the move away from monolithic applications toward smaller, reusable services with discrete functionality. such services can be combined and recombined to deliver different applications to users. web services is an implementation of soa principles. the term describes the use of technologies such as xml to enable the seam- less interoperability of web-based applications. ajax enables web services and hence enables soa principles. thus, the adoption of ajax facilitates the move toward soa and all the advantages of reuse and integration that this offers. ■ arc arc is an experimental open-source metasearch pack- age available for download from the sourceforge open- source foundry. it can be configured to harvest open using ajax to empower dynamic searching | wusteman archives initiative-protocol for metadata harvesting (oai-pmh)-compliant data from multiple repositories. the harvested results are stored in a relational database and can be searched using basic web forms. arc’s advanced search form is illustrated in figure . ■ applying ajax to the search gui the use of ajax has the potential to narrow the gulf between the responsiveness of guis for web applications and those for desktop applications. the flexibility, usabil- ity, and richness of the latter are now possible for the former. the ojax gui, illustrated in figure , has been developed to demonstrate how ajax can improve the richness of arc-like guis. ojax, including full source code, is available under the open-source apache license and is hosted on sourceforge. ojax comprises a client-side gui, implemented in javascript and html, and server-side metasearch web services, implemented in java. the web services connect directly to a metasearch database created by arc from harvested repositories. the database connectivity lever- ages several libraries from the apache jakarta project, which provides open-source java solutions. ■ development process the ojax gui was developed iteratively using agile software development methods. features were added incrementally and feedback gained from a proxy user. in order to gain an in-depth understanding of the sys- tem and the implications for the remainder of the gui, features were initially built from scratch, using object- oriented javascript.they were then rebuilt using three open-source javascript libraries: prototype, script.aculo .us, and rico. prototype provides base ajax capability. it also includes advanced functionality for object-oriented javascript, such as multiple inheritance. the other two libraries are built on top of prototype. the script.aculo. us library specializes in dynamic effects, such as those used in auto-completion. the rico library, developed by sabre, provides other key javascript effects—for example, dynamic scrollable areas and dynamic sorting. ■ storyboard one of the aims of the national information standards organization (niso) metasearch initiative is to enable all library users to “enjoy the same easy searching found in web-based services like google.” adopting this approach, ojax incorporates the increasingly common concept of the search bar, popularized by the google toolbar. ojax aims to be as simple, uncluttered, and unthreatening as possible. the goal is to reflect the sim- ple-search experience while, at the same time, providing the power of an advanced search. thus, the user interface has been kept as simple as possible while maintaining equivalent functionality with the arc advanced search interface. all arc functionality, with the exception of the grouping feature, is provided. to help the intuitive flow of the operation, the fields are set out as a sentence: find [term(s)] in [all archives] from [earliest year] until [this year] in [all subjects] tool tips are available for text-entry fields. by default, searching is on author, title, and abstract. these fields map to the creator, title, and description dublin core meta- data fields harvested from the original repositories. the search can be restricted by deselecting unwanted fields. arc supports both mysql and oracle databases. mysql has been chosen for ojax as mysql is an open-source database. boolean search syntax has been figure . arc’s advanced search form figure . the ojax metasearch user interface information technology and libraries | june implemented in ojax to allow for more powerful search- ing. the syntax is similar to that used by google in that it identifies and/or and exact phrase functionality by +/- and “ ”. hence it preserves the user’s familiarity with basic google search syntax. however, it is not as powerful as the full google search syntax; for example, it does not support query modifiers such as: intitle: the focus of this research is the application of ajax to the search gui and not the optimization of the power or expressive capability of the underlying search engine. however, the implementation of an alternative back end that uses a full-text search engine, such as apache lucene, would improve the expressive power of advanced que- ries. full-text search expressiveness is likely to be key to the usability of ojax, ensuring its adequacy for the advanced user without alienating the novice. ■ unifying the user interface one of the main aims of ojax is the unification of the user interface. instead of offering distinct options for simple and advanced search and for refining a completed search, the interface is sufficiently dynamic to make this unnecessary. the user need never navigate between pages because all options, both simple and advanced, are available from the same page. and all results are made available on that same page in the form of a scrollable list. the only point at which a new page is presented is when the resource identifier of a result is clicked. at this stage, a pop-up window, external to the ojax session, displays the full metadata for that resource. this page is generated by the external repository from which the record was originally harvested. simple and advanced search options are usually kept separate because most users are unwilling or unable to use the latter. furthermore, the design of existing search-user interfaces is based on the assumption that the retrieval of results will be sufficiently time-consuming that users will want to have selected all options beforehand. with ojax, however, users do not have to make a complete choice of all the options they might want to try before they see any results. as data are entered, answers flow to accommodate them. because the inter- face is so dynamic and responsive and because users are given immediate feedback, they do not have to be concerned about wasting time due to the wrong choice of search options. users iterate toward the search results they require by manipulating the results in real time. the reduced level of investment that users must make before they achieve any return from the system should encourage them to experiment, hence promoting enac- tive learning. ■ auto-completion in order to provide instant feedback to the user, the search-terms field and the subject field use ajax to auto- complete user entries. figure illustrates the result of typing smith in the search-terms field. a list is automati- cally dropped down that itemizes all matches and the number of their occurrences. users select the term they want, the entire field is automatically completed, and a search is triggered. the arc system denormalizes some of the harvested data before saving them in its database. for example, it merges all the author fields into one single field, each name separated by a bar character. to enable the ojax auto-completion feature, it was necessary to renormalize the names. a new table is used to store each name in a separate row; names are referenced by the resource iden- tifier. to enable this, arc’s indexing code was updated so that it creates this table as it indexes records extracted from the oai-pmh feed. in its initial implementation, ojax uses a simple algorithm for auto-completion. future work will involve developing a more complex heuristic that will return results more closely satisfying user requirements. ■ auto-search as already mentioned, a central theme of ojax is the attempt to reduce the commitment necessary from users before they receive feedback on their actions. one way in which dynamic feedback is provided is the triggering of an immediate search whenever an entire option has been selected. examples of entire options include choice of an archive or year and acceptance of a suggested auto- completion. in addition, the following heuristics are used to identify when a user is likely to have finished entering a search term and, thus, when a search should be triggered: . entering a space character in the search-terms field or subject field . tabbing out of a field after having modified its con- tents . five seconds of user inactivity for a modified field the third heuristic aims to catch some of the edge cases that the other heuristics may miss. it is assumed likely that a term has been completed if a user has made no edits in the last five seconds. as each term will be using ajax to empower dynamic searching | wusteman separated by a space, it is only the last term in a search phrase that is likely not to trigger an auto-search via the first heuristic. users can click the search button whenever they wish, but they should never have to click it. the zuggest sys- tem abandons the search button entirely; ojax retains it, mainly in order to avoid confounding user expectations. while a search is in progress, the search button is greyed out and acquires a red border. this is particularly useful in alerting the user that a search has been auto- matically triggered. this is the only feature of ojax that may have an impact on network load in terms of slightly higher traffic. however, the increased number of requests is offset by a reduction in the size of each response because the gui is not downloaded with it. for example, initiating a search in arc results in an average response size of . k. the response is in the form of a complete html page. initiating a search in ojax results in an average response size of . k. the latter comprises a web service response in xml. in other words, more than seven ojax auto- searches would have to be triggered before the size of the initial search result in arc was exceeded. ■ dynamic archive list the use of ajax enables a static html page to contain a small component of dynamic data without the entire page having to be dynamically generated on the server. ojax illustrates this: the contents of the drop-down box listing the searchable archives are not hard-coded in the html page. rather, when the page is loaded, an ajax request for the set of available archives is generated. this is a useful technique; static html pages can be cached by browsers and proxy servers, and only the dynamic portion of the data, perhaps those used to personalize the page, need be downloaded at the start of a new session. ■ dynamic scrolling searches commonly produce thousands of results. typ- ical systems, such as google and arc, make these results available via a succession of separate pages, thus requiring users to navigate between them. finding information by navigating multiple pages can take longer than scrolling down a single page, and users rarely look beyond the second page of search results. to avoid these problems and to encourage users to look at more of the available results, those results could be made available in one scrollable list. but, in a typical non-ajax application, accessing a scrollable list of, say, two thousand items would require the entire list to be downloaded via one enormous html page. this would be a huge operation; if it did not crash the browser, it would, at least, result in a substantial wait for the user. the rico library provides a feature to enable dynamic scrollable areas. it uses ajax to fetch more records from the server when the user begins to scroll off the visible area. this is used in the display of search results in ojax, as illustrated in figure . to the user, it appears that the scrollable list is seamless and that all , search results are already downloaded. in fact, only have been downloaded. the rest are available at the server. as the user scrolls further down, say to item , an ajax request is made for the next ten items. any item downloaded is cached by the ajax engine and need not be requested again if, for example, the user scrolls back up the list. a dynamic information panel is available to the right of the scroll bar. it shows the current scroll position in relation to the beginning and end of the results set. in figure . auto-completion in the search terms field figure . display of search results and dynamic information panel information technology and libraries | june figure , the information panel indicates that there are , results for this particular search and that the cur- rent scroll position is at result number . this number updates instantly during scrolling, preserving the illusion that all results have been downloaded and providing users with dynamic feedback on their progress through the results set. this means that users do not have to wait for the main results window to refresh to identify their current position. ■ auto-expansion of results ojax aims to provide a compact display of key informa- tion, enabling users to see multiple results simultane- ously. it also aims to provide simple access to full result details without requiring navigation to a new web page. in the initial results display, only one line each of the title, authors, and subject fields, and two lines of the abstract, are shown for each item. as the cursor is placed on the relevant field, the display expands to show any hidden detail in that field. at the same time, the back- ground color of the field changes to blue. when the cur- sor is placed on the bar containing the resource identifier, all display fields for that item are expanded, as illustrated in figure . this expansion is enabled via simple cascading style sheet (css) features. for example, the following css dec- laration hides all but the first line of authors: #searchresults td div { overflow:hidden; height: . em } when the cursor is placed on the author details, the overflow becomes visible and the display field changes its dimensions to fit the text inside it: #searchresults td div:hover { overflow:visible; height:auto } ■ sorting results another method used by ojax to minimize upfront user investment is to provide initial search results before requiring the user to decide on sort options. because results are available so quickly and because they can be re-sorted so rapidly, it is not necessary to offer pre-search selection of sort options. ajax facilitates rapid presen- tation of results; after a re-sort, only those on the first screen must be downloaded before they can be presented to the user. results may be sorted by title, author, subject, abstract, and resource identifier. these options are listed on the gray bar immediately above the results list. clicking one of these options sorts the results in ascending order; an upward-pointing arrow appears to the right of the sort option chosen, as illustrated in fig- ure . clicking on the option again sorts in descending order and reverses the direction of the arrow. clicking on the arrow removes the sort; the results revert to their original order. functionality for the sort feature is provided by the rico javascript library. server-side implementation sup- ports these features by caching search results so that it is not necessary to regenerate them via a database query each time. figure . auto-expansion of all fields for item number figure . results being sorted in ascending order by title using ajax to empower dynamic searching | wusteman ■ search history several experimental systems—for example, zuggest— have employed ajax to facilitate a search-history feature. a similar feature could be provided for ojax. a button could be added to the right of the results list. when cho- sen, it could expand a collapsible search-history sidebar. as the cursor was placed on one of the previous searches listed in the sidebar, a call out, that is, a speech bubble, could be displayed. this could provide further informa- tion such as the number of matches for that search and a summary of the search results clicked on by the user. clicking one of the previous searches would restore those search results to the main results window. this feature would take advantage of the ajax per- sistent javascript engine to maintain the history. its use could help counter concerns about ajax technology “breaking” the back button; the feature could be imple- mented so that the back button returned the user to the previous entry in the search history. in fact, this imple- mentation of back-button functionality could be more useful than the implementation in google, where hitting the back button is likely to take the user to an interim results page; for example, it might simply take the user from page of results to page of results. ■ scrapbook users browsing through search results on ojax would require some simple method of maintaining a record of those resource details that interested them. ajax could enable the development of a useful scrapbook feature to which such resource details could be copied and stored in the persistent javascript engine. ojax could further leverage a shared bookmark web service, such as del. icio.us or furl, to save the scrapbook for use in future ses- sions and to share it with other members of a research or interest group. ■ potential developments for ojax as well as searching a database of harvested metadata, the ojax user interface could also be used to search an oai-pmh-compliant repository directly. with appropri- ate implementation, all of ojax’s current features could be made available, apart from auto-completion. a recent development has enabled the direct indexing of repositories by google using oai-pmh. the latter provides google with additional metadata that can be searched via the google web services apis. the current ojax web services could be replaced by the google apis, thus eliminating the need for ojax to host any server-side components. hence, ojax could become an alternative gui for google searching. ■ conclusion ojax demonstrates that the use of ajax can enable features in web applications that, until now, have been restricted to desktop applications. in ojax, it facilitates a simple, nonthreatening, but powerful search user inter- face. page navigation is eliminated; dynamic feedback and a low initial investment on the part of users encour- age experimentation and enable enactive learning. the use of ajax could similarly transform other web applica- tions aimed at library patrons. however, ajax is still maturing, and the barrier to entry for developers remains high. we are a long way from an ajax button appearing in dreamweaver. reusable, well-tested components, such as rico, and software frameworks, such as ruby on rails, sun’s j ee framework, and microsoft’s atlas, will help to make ajax technology accessible to a wider range of developers. as with all new technologies, there is a temptation to use ajax simply because it exists. as ajax matures, it is important that its focus does not become the enabling of “cool” features but remains the optimization of the user experience. references and notes . ojax homepage, http://ojax.sourceforge.net (accessed apr. , ). . j. j. garrett, “ajax: a new approach to web applica- tions,” feb. , , www.adaptivepath.com/publications/ essays/archives/ .php (accessed nov. , ). . ibid. . j. nielsen, “the need for speed,” alertbox mar. , , www.useit.com/alertbox/ a.html (accessed nov. , ). . dynamic html and xml: the xmlhttprequest object, http://developer.apple.com/internet/webcontent/xmlhttpreq .html (accessed apr. , ). . javascript object notation, wikipedia definition, http:// en.wikipedia.org/wiki/json (accessed apr. , ). . google gmail, http://mail.google.com (accessed apr. , ); google suggest, www.google.com/webhp?complete = &hl=en (accessed apr. , ); google groups, http://groups .google.com (accessed apr. , ); google maps, http://maps .google.com (accessed apr. , ). . p. binkley, “ajax and auto-completion,” quædam cuiusdam blog may , , www.wallandbinkley.com/quaedam/?p= (accessed nov. , ). . francis shanahan, zuggest, www.francisshanahan.com/ zuggest.aspx (accessed apr. , ). information technology and libraries | june . a. rhyno, “ajax and the rich web inter- face,” librarycog blog apr. , , http://librarycog .uwindsor.ca: /artblog/librarycog/ (accessed nov. , ); r. tennant, “tennant’s top tech trend tidbit,” lita blog june , , http://litablog.org/?p= (accessed nov. , ). . t. hickey, “ajax and web interfaces,” outgoing blog, mar. , . retrieved nov. , http://outgoing.typepad .com/outgoing/ / /web_application.html. . oclc deweybrowser. http://ddcresearch.oclc.org/ ebooks/fileserver (accessed apr. , ). . hickey, “ajax and web interfaces.” . j. wusteman, “from ghostbusters to libraries: the power of xul,” library hi tech , no ( a). retrieved nov. , www.ucd.ie/wusteman/; cover pages, microsoft exten- sible application markup language (xaml), http://xml.cover pages.org/ms-xaml.html (accessed apr. , ). . google extensions for firefox, http://toolbar.google .com/firefox/extensions/index.html (accessed apr. , ). . c. adams, “ajax: usable interactivity with remote scripting,” sitepoint. (jul. , ), www.sitepoint.com/article/ remote-scripting-ajax (accessed nov. , ). . xhtml . , w c working draft, may , , www .w .org/tr/ /wd-xhtml - (accessed apr. , ). . client/server model, http://en.wikipedia.org/wiki/ client/server (accessed apr. , ). . dll hell, http://en.wikipedia.org/wiki/dll_hell (accessed apr. , ). . j. nielsen, “the need for speed.” . service-oriented architecture, http://en.wikipedia.org/ wiki/service-oriented_architecture (accessed apr. , ). . j. wusteman, “realizing the potential of web services,” oclc systems & services: international digital library perspectives , no. ( ): – . . arc—a cross archive search service, old dominion university digital library research group, http://arc.cs.odu .edu (accessed apr. , ); niso metasearch initiative, www .niso.org/committees/ms_initiative.html (accessed apr. , ); arc download page, sourceforge, http://oaiarc.source forge.net (accessed apr. , ). . open archives initiative protocol for metadata harvest- ing, www.openarchives.org/oai/openarchivesprotocol.html (accessed apr. , ). . ojax download page, sourceforge, http://sourceforge .net/projects/ojax (accessed apr. , ). . apache jakarta project, http://jakarta.apache.org (accessed apr. , ); apache jakarta commons dbcp, http:// jakarta.apache.org/commons/dbcp (accessed apr. , ); apache jakarta commons dbutils, http://jakarta.apache.org/ commons/dbutils (accessed apr. , ). . agile software development definition, wikipedia, http://en.wikipedia.org/wiki/agile_software_development (accessed apr. , ). . prototype javascript framework, http://prototype.conio .net (accessed apr. , ); script.aculo.us, http://script.aculo .us (accessed apr. , ); rico, http://openrico.org/rico/ home.page (accessed apr. , ). . sabre, www.sabre.com (accessed apr. , ). . niso metasearch initiative, www.niso.org/committees/ ms_initiative.html (accessed apr. , ). . google toolbar, http://toolbar.google.com (accessed apr. , ). . dublin core metadata initiative, http://dublincore.org (accessed apr. , ). . mysql, www.mysql.com (accessed apr. , ). . google help center, advanced operators, www.google .com/help/operators.html (accessed apr. , ). . apache lucene, http://lucene.apache.org (accessed apr. , ). . j. nielsen, “search: visible and simple,” alertbox may , , www.useit.com/alertbox/ .html (accessed nov. , ). . francis shanahan, zuggest. . j. r. baker, “the impact of paging versus scrolling on reading online text passages,” usability news , no. ( ), http://psychology.wichita.edu/surl/usabilitynews/ / paging_scrolling.htm (accessed nov. , ); j. nielsen, “search: visible and simple.” . j. j. garrett, “ajax: a new approach to web applica- tions.” . del.icio.us, http://del.icio.us (accessed apr. , ); furl, www.furl.net (accessed apr. , ). . google sitemaps (beta) help, www.google.com/web masters/sitemaps/docs/en/other.html (accessed apr. , ). . ruby on rails, www.rubyonrails.org (accessed apr. , ); java platform, enterprise edition (j ee), http://java .sun.com/j ee (accessed apr. , ); m. lamonica, “microsoft gets hip to ajax,” cnet news.com, june , , http:// news.com.com/microsoft+gets+hip+to+ajax/ - _ - .html (accessed nov. , ). information technology and libraries | june author name and second author author id box for column layout this paper discusses google scholar as an extension of kilgour’s goal to improve the availability of information. kilgour was instrumental in the early development of the online library catalog, and he proposed passage retrieval to aid in information seeking. google scholar is a direct descendent of these technologies foreseen by kilgour. google scholar holds promise as a means for libraries to expand their reach to new user communities, and to enable libraries to provide quality resources to users dur- ing their online search process. editor’s note: this article was submitted in honor of the fortieth anniversaries of lita and ital. f red kilgour would probably approve of google scholar. kilgour wrote that the paramount goal of his professional career is “improving the availabil- ity of information.” he wrote about his goal of achieving this increase through shared electronic cataloging, and even argued that shared electronic cataloging will move libraries toward the goal of percent availability of information. throughout much of kilgour’s life, percent avail- ability of information meant that all of a library’s books would be on the shelves when a user needed them. in proposing shared electronic cataloging—in other words, online union catalogs—kilgour was proposing that users could identify libraries’ holdings without having to travel to the library to use the card catalog. this would make the holdings of remote libraries as visible to users as the holdings of their local library. kilgour went further than this, however, and also pro- posed that the full text of books could be made available to users electronically. this would move libraries toward the goal of percent availability of information even more than online union catalogs. an electronic resource, unlike physical items, is never checked out; it may, in theory, be simultaneously used by an unlimited number of users. where there are restrictions on the number of users of an electronic resource—as with subscription ser- vices such as netlibrary, for example—this is not a neces- sary limitation of the technology, but rather a limitation imposed by licensing and legal arrangements. kilgour understood that his goal of percent availability of information would only be reached by leveraging increasingly powerful technologies. the exis- tence of effective search tools and the usability of those tools would be crucial so that the user would be able to locate available information without assistance. to achieve this goal, therefore, kilgour proposed and was instrumental in the early development of much library automation: he was behind the first uses of punched cards for keeping circulation records, he was behind the development of the first online union catalog, and he called for passage retrieval for information seeking at a time when such systems were first being developed. this development and application of technology was all directed toward the goal of improving the availability of information. kilgour stated that the goal of these pro- posed information-retrieval and other systems was “to supply the user with the information he requires, and only that information.” shared catalogs and electronically available text have the effect of removing both spatial and temporal barriers between the user and the material being used. when the user can access materials “from a personal microcom- puter that may be located in a home, dormitory, office, or school,” the user no longer has to physically go to the library. this is a spatial barrier when the library is located at some distance from the user, or if the user is physically constrained in some way. even if the user is perfectly able-bodied, however, and located close to a library, electronic access still eliminates a temporal bar- rier: accessing materials online is frequently faster and more convenient than physically going to the library. electronic access enables percent availability of information in two ways: by ensuring that the material is available when the user wants it, and by lowering or removing any actual or perceived barriers to the user accessing the material. ■ library automation weise writes that “for at least the last twenty to thirty years, we [librarians] have done our best to provide them [users] with services so they won’t have to come to the library.” the services that weise is referring to are the ability for users to search for and gain access to the full text of materials online. libraries of all types have widely adopted these services: for example, at the author’s own institution, the university of north carolina at chapel hill, the libraries have subscriptions to approximately seven hundred databases and provide access to more than , unique periodical titles; many of these sub- scriptions provide access to the full text of materials. additionally, the state library of north carolina pro- vides a set of more than one hundred database subscrip- tions to all academic and public libraries around the jeffrey pomerantz jeffrey pomerantz (pomerantz@unc.edu) is assistant pro- fessor in the school of information and library science, university of north carolina at chapel hill. google scholar and percent availability of information google scholar and percent availability of information | pomerantz state; any north carolina resident with a library card may access these databases. several other states have similar programs. by providing users with remote access to materials, libraries have created an environment in which it is possible for users to be remote from the library. or rather, as lipow points out, it is the library that is remote from the user, yet the user is able to seek and find information. this adoption of technology by libraries has had the effect of enabling and empowering users to seek informa- tion for themselves, without either physically going to a library or seeking a librarian’s assistance. the increasing sophistication of freely available tools for information seeking on the web has accelerated this trend. in many cases, users may seek information for themselves online without making any use of a library’s human-intermedi- ated or other traditional services. (certainly, providing access to electronic collections may be considered to be a service of the library, but this is a service that may not require the user either to be physically in the library or to communicate with a librarian.) even technically unsophisticated users may use a search engine and locate information that is “good enough” to fulfill their infor- mation needs, even if it is not the ideal or most complete information for those purposes. thus, for better or worse, the physical library is no longer the primary focus for many information seekers. part of this movement by users toward self-sufficiency in information seeking is due to the success of the web search engine, and to the success of google in particular. recent reports from the pew internet and american life project shed a great deal of light on users’ use of these tools. rainie and horrigan found that “on a typical day at the end of , some million american adults logged onto the internet.” fallows found that “on any given day, % of those online use search engines.” fallows, rainie, and mudd found that of their respondents, “ % say that google is their top choice of search engine.” from these figures, it can be roughly estimated that more than mil- lion people use search engines, and more than million use google on any given day—and that is only within the united states. this trend seems quite dark for libraries, but it actu- ally has its bright side. it is important to make a distinc- tion here between use of a search engine and use of a reference service or other library service. there is some evidence that users’ questions to library reference ser- vices are becoming more complex. why this is occur- ring is less clear, but it may be hypothesized that users are locating information that is good enough to answer their own simple questions using search engines or other internet-based tools. the definition of “good enough” may differ considerably between a user and a librarian. nevertheless, one function of the library is education, and as with all education, the ultimate goal is to make the student self-sufficient in self-teaching. in the context of the library, this means that one goal is to make the user self-sufficient in finding, evaluating, and using informa- tion resources. if users are answering their own simple questions, and asking the more difficult questions, then it may be hypothesized that the widespread use of search engines has had a role in raising the level of debate, so to speak, in libraries. rather than providing instruction to users on simply using search engines, librarians may now assume that some percentage of library users possess this skill, and may focus on teaching higher-level infor- mation-literacy skills to users (www.ala.org/ala/acrl/ acrlstandards/informationliteracycompetency.htm). simple questions that users may answer for them- selves using a search engine, and complex questions requiring a librarian’s assistance to answer are not oppo- sites, of course, but rather two ends of a spectrum of the complexity of questions. while the advance of online search tools may enable users to seek and find informa- tion for themselves at one end of this spectrum, it seems unlikely that such tools will enable users to do the same across the entire spectrum any time soon; perhaps ever. the author believes that there will continue to be a role for librarians in assisting users to find, evaluate, and use information. it is also important to make another distinction here, between the discovery of resources, and access to those resources. libraries have always provided mechanisms for users to both discover and access resources. neither the card catalog nor the online catalog contains the full text of the materials cataloged; rather, these tools are means to enable the user to discover the existence of resources. the user may then access these resources by visiting the library. search engines, similar to the card and online catalogs, are tools primarily for discovery of resources: search-engine databases may contain cached copies of web pages, but the original (and most up-to- date) version of the web page resides elsewhere on the web. thus, a search engine enables the user to discover the existence of web pages, but the user must then access those web pages elsewhere. the author believes that there will continue to be a role for libraries in providing access to resources—regardless of where the user has dis- covered those resources. in order to ensure that libraries and librarians remain a critical part of the user’s information-seeking process, however, libraries must reappropriate technologies for online information seeking. search engines may exist separate from libraries, and users may use them without making use of any library service. however, libraries are already the venue through which users access much online content—newspapers, journals, and other peri- odicals; reference sources; genealogical materials—even if many users do not physically come to the library or consult a librarian when using them. it is possible for information technology and libraries | june libraries to add value to search technologies by providing a layer of service available to those using it. ■ google scholar one such technology for online information seeking to which libraries are already adding value, and that could add value to libraries in turn, is google scholar (scholar. google.com). google scholar is a specialty search tool, obviously provided by google, which enables the user to search for scholarly literature online. this literature may be on the free web (as open-access publications become more common and as scholars increasingly post preprint or post-print copies of their work on their personal web sites), or it may be in subscription databases. users may access literature in subscription databases in one of two ways: ( ) if the user is affiliated with an institution that subscribes to the database, the user may access it via whatever authentication method is in place at the institu- tion (e.g., ip authentication, a proxy server), or ( ) if the user is not affiliated with such an institution, the user may pay for access to individual resources on a pay-per- view basis. there is not sufficient space here to explore the details of google scholar’s operation, and anyway that is not the point of this paper; for excellent discussions of the operation of google scholar, see gardner and eng, and jacsó. pace draws a distinction between federated searching and metasearching: federated search tools compile and index all resources proactively, prior to any user’s actual search, in a just-in-case approach to users’ searching. metasearch tools, on the other hand, search all resources on the fly at the time of a user’s search, in a just-in-time approach to users’ searching. google scholar is a feder- ated search tool—as, indeed, are all of google’s current services—in that the database that the user searches is compiled prior to the user’s actual search. in this, google scholar is a direct descendent of kilgour’s work to develop shared online library catalogs. a shared library catalog is a union catalog: it is a database of libraries’ physical holdings, compiled prior to any actual user’s search. google scholar is also a union catalog, though a catalog of publishers’ electronic offerings pro- vided by libraries, rather than of libraries’ physical hold- ings. it should be noted, however, that while this difference is an important one for libraries and publishers, it might not be understood or even relevant for many users. many of the resources indexed in google scholar are also available in full text. this fact allows google scholar to also move in the direction of kilgour’s goal of making passage retrieval possible for scholarly work. by using google’s core technology—the search engine and the inverted index that is created when pages are indexed by a search engine—google scholar enables full-text search- ing of scholarly work. as mentioned above, when users search google scholar, they retrieve a set of links to the scholarly literature retrieved by the search. google scholar also makes use of google’s link- analysis algorithms to analyze the network of citations between publications—instead of the network of hyper- links between web pages, as google’s search engine more typically analyzes. a cited by link is included with each retrieved link in google scholar, stating how many other publications cite the publication listed. clicking on this cited by link performs a preformulated search for those publications. this citation-analysis functionality resembles the functionality of one of the most common and widely used scholarly databases in the scholarly com- munity: the isi web of science (wos) database (scientific .thomson.com/products/wos). wos enables users to track citations between publications. this functionality has wide use in scholarly research, but until google scholar, it has been largely unknown outside of the scholarly community. with the advent of google scholar, however, this functionality may be employed by any user for any research. further, there is a plugin for the firefox browser (www.mozilla.com/firefox) that displays an icon for every record on the page of retrieved results that links to the appropriate record in the library’s opac (google scholar does not, however, currently provide this func- tionality natively ). this provides a link from google scholar to the materials that the library holds in its col- lection. when the item is a book, for example, this link to the opac enables users to find the call number of the book in their local library. when the item is a journal, it enables them to find both the call number and any data- base subscriptions that index that journal title. periodicals are often indexed in multiple databases, so libraries with multiple-database subscriptions often have multiple means of accessing electronic versions of journal titles. a library user may access a periodical via any or all of these individual subscriptions without using google scholar— but to do so, the user must know which database to use, which means knowing either the topical scope of a data- base or knowing which specific journals are indexed in a database. as a more centralized means of accessing this material, many users may prefer a link in google scholar to the library’s opac. google scholar thus fulfills, in large part, kilgour’s vision of shared electronic cataloging. in turn, shared cata- loging goes a long way toward achieving kilgour’s vision of percent availability of information by allowing a user to discover the existence of information resources. however, discovery of resources is only half of the equa- tion: the other half is access to those resources. and it is here where libraries may position themselves as a critical part of the information-seeking process. search engines google scholar and percent availability of information | pomerantz may enable users to discover information resources on their own, without making use of a library’s services, but it is the library that provides the “last mile” of service, enabling users to gain access to many of those resources. ■ conclusion google scholar is the topic of a great deal of debate, both in the library arena and elsewhere. unlike union catalogs and many other online resources used in librar- ies, it is unknown what materials are included in google scholar, since as of this writing google has not released information about which publishers, titles, and dates are indexed. google is known to engage in self-censor- ship—or self-filtering, depending on what coverage one reads—and so potentially conflicts with the american library association’s freedom to read statement (www .ala.org/ala/oif/statementspols/ftrstatement/freedom readstatement.htm). google is a commercial entity and, as such, a primary motivation of google must be profit, and only secondarily, meeting the information needs of library users. for all of these and other reasons, there is considerable debate among librarians about whether it is appropriate for libraries to provide access to google scholar. despite this debate, however, users are using google scholar. google scholar is simply the latest tool to enable users to seek information for themselves; it isn’t the first and it won’t be the last. google scholar holds a great deal of promise for libraries due to the combination of google’s popularity and ease of use, and the resources held by or subscribed to by libraries to which google scholar points. as kesselman and watstein suggest, “libraries and librarians need to have a voice” in how tools such as google scholar are used, given that “we are the ones most passionate about meeting the information needs of our users.” given that library users are using google scholar, it is to libraries’ benefit to see that it is used well. google scholar is the latest tool in a long history of information-seeking technologies that increasingly real- ize kilgour’s goal of achieving percent availability of information. google scholar does not provide access to percent of information resources in existence; but rather enables discovery of information resources, and allows for the possibility that these resources will be dis- coverable by the user percent of the time. google scholar may be on the vanguard of a new way of integrating library services into users’ everyday information-seeking habits. as taylor tells us, people have their own individual sources to which they go to find information, and libraries—for many people—are not at the top of their lists. google, however, is at the top of the list for a great many people. properly harnessed by libraries, therefore, google scholar has the potential to bring users to library resources when they are seeking information. google scholar may not bring users physically to the library. instead, what google scholar can do is bring users into contact with resources provided by the library. this is an important distinction, because it reinforces a change that libraries have been undergoing since the advent of the online database: that of providing access to materials that the library may not own. ownership of materials potentially allows for a greater measure of con- trol over the materials and their use. ownership in the context of libraries has traditionally meant ownership of physical materials, and physical materials by nature restrict use, since the user must be physically collocated with the materials, and use of materials by one user precludes use of those materials by other users for the duration of the use. providing access to materials, on the other hand, means that the library may have less control over materials and their use, but this potentially allows for wider use of these materials. by enabling users to come into contact with library resources in the course of their ordinary web searches, google scholar has the potential to ensure that libraries remain a critical part of the user’s information-seeking process. it benefits google when a library participates with google scholar, but it also benefits the library and the library’s users: the library is able to provide users with a familiar and easy-to-use path to materials. this is (for lack of a better term) a “spoonful of sugar” approach to seeking and finding information resources: by using an interface that is familiar to users, libraries may provide quality information sources in response to users’ informa- tion seeking. green wrote that “a librarian should be as unwilling to allow an inquirer to leave the library with his ques- tion unanswered as a shop-keeper is to have a customer go out of his store without making a purchase.” a modern version of this might be that a librarian should be as unwilling to allow an inquirer to abandon a search with his question unanswered. google scholar and online tools like it have the potential to draw users away from libraries; however, these tools also have the potential to usher in a new era of service for libraries: an expansion of the reach of libraries to new users and user communities; a closer integration with users’ searches for information; and the provision of quality resources to all users, in response to all information needs. google scholar and online tools like it have the potential to enable libraries to realize kilgour ’s goals of improv- ing the availability of information, and to provide percent availability of information. these are goals on which all libraries can agree. information technology and libraries | june ■ acknowledgements many thanks to lisa norberg, instruction librarian, and timothy shearer, systems librarian, both at the university of north carolina at chapel hill, for many extensive conversations about google scholar, which approached coauthorship of this paper. this paper is dedicated to the memory of kenneth d. shearer. references and notes . frederick g. kilgour, “historical note: a personalized prehistory of oclc,” journal of the american society for informa- tion science , no. ( ): . . frederick g. kilgour, “future of library computerization,” in current trends in library automation: papers presented at a work- shop sponsored by the urban libraries council in cooperation with the cleveland public library, alex ladenson, ed. (chicago: urban libraries council, ), – ; frederick g. kilgour, “toward percent availability,” library journal , no. ( ): – . . kilgour, “toward percent availability.” . frederick g. kilgour, “lack of indexes in works on infor- mation science,” journal of the american society for information science , no. ( ): ; frederick g. kilgour, “implications for the future of reference/information service,” in collected papers of frederick g. kilgour: oclc years, lois l. yoakam, ed. (dublin, ohio: oclc online computer library center, inc., ): – . . frederick g. kilgour, “a new punched card for circula- tion records,” library journal , no. ( ): – ; kilgour, “historical note”; frederick g. kilgour and nancy l. feder, “quotations referenced in scholarly monographs,” journal of the american society for information science , no. ( ): – ; gerald salton, j. allan, and chris buckley, “approaches to pas- sage retrieval in full-text information systems,” in proceedings of the th annual international acm sigir conference on research and development in information retrieval (new york: acm pr., ), – . . kilgour, “implications for the future of reference/infor- mation service,” . . kilgour, “toward percent availability,” . . frieda weise, “being there: the library as place,” journal of the medical library association , no. ( ): , www.pubmedcentral.nih.gov/articlerender.fcgi?artid= (accessed apr. , ). . it is difficult to determine precise figures, as there is con- siderable overlap in coverage; several vendors provide access to some of the same periodicals. . north carolina’s database subscriptions are via the nc live service, www.nclive.org (accessed apr. , ). . anne g. lipow, “serving the remote user: reference service in the digital environment,” paper presented at the ninth australasian information online and on disc conference and exhi- bition, sydney, australia, – jan. , www.csu.edu.au/ special/online /proceedings / .htm (accessed apr. , ). . j. janes, “academic reference: playing to our strengths,” portal: libraries and the academy , no. ( ): – , http:// muse.jhu.edu/journals/portal_libraries_and_the_academy/ v / . janes.html (accessed apr. , ). . lee rainie and john horrigan, a decade of adoption: how the internet has woven itself into american life (washington, d.c.: pew internet & american life project, ), , www.pewinter net.org/ppf/r/ /report_display.asp (accessed apr. , ). . deborah fallows, search engine users (washington, d.c.: pew internet & american life project, ), i, www.pew internet.org/pdfs/pip_searchengine_users.pdf (accessed apr. , ). . deborah fallows, lee rainie, and graham mudd, data memo on search engines (washington, d.c.: pew internet & american life project, ), , www.pewinternet.org/ppf/ r/ /report_display.asp (accessed apr. , ). . laura bushallow-wilber, gemma devinney, and fritz whitcomb, “electronic mail reference service: a study,” rq , no. ( ): – ; carol tenopir and lisa a. ennis, “reference services in the new millennium,” online , no. ( ): – . . alma swan and sheridan brown, open access self- archiving: an author study (truro, england: key perspectives, ), www.jisc.ac.uk/uploaded_documents/open% access % self% archiving-an% author% study.pdf (accessed apr. , ). . susan gardner and susanna eng, “gaga over google? scholar in the social sciences,” library hi tech news ( ): – ; péter jacsó, “google scholar: the pros and the cons,” online information review , no. ( ): – . . andrew pace, “introduction to metasearch . . . and the niso metasearch initiative,” presentation to the openurl and metasearch workshop, sept. – , , www.niso.org/news/ events_workshops/openurl- -ppts/ - -pace.ppt (accessed apr. , ). . this plugin was developed by peter binkley, digital ini- tiatives technology librarian at the university of alberta. see www.ualberta.ca/~pbinkley/gso (accessed apr. , ). . see, for example, gardner and eng, “gaga over google?”; jacsó, “google scholar”; m. kesselman and s. b. watstein, “google scholar and libraries: point/counterpoint,” reference services review , no. ( ): – . . jacsó, “google scholar.” . anonymous, google censors itself for china, bbc news, jan. , , http://news.bbc.co.uk/ /hi/technology/ .stm (accessed apr. , ); a. mclaughlin, “google in china,” google blog., jan. , , http://googleblog.blogspot .com/ / /google-in-china.html (accessed apr. , ). . kesselman and s. b. watstein, “google scholar and libraries,” . . robert s. taylor, “question-negotiation and information seeking in libraries,” college & research libraries , no. ( ): – . . fallows, rainie, and mudd, data memo on search engines. . samuel s. green, “personal relations between librarians and readers,” american library journal i, no. – ( ): . – . digitization has bestowed upon librarians and archivists of the late th and early st centuries the opportunity to reexamine how they access their collections. it draws these two traditional groups together with it specialists in order to collaborate on this new great challenge. in this paper, the authors offer a strategy for adapting a library system to traditional archival practice. t he librarian and the archivist . . . both collect, pre- serve, and make accessible materials for research; but significant differences exist in the way these materials are arranged, described, and used.” among the items usually collected by libraries are: published books and serials, and in more recent times, commercially available sound recordings, films, videos, and electronic resources of various types. archives, on the other hand, tend to collect original records of an organization, unique personal papers, as well as other effects of individuals and families. each type of institution, given its particular emphasis, has its own traditions and its own methods of dealing with its collections. most mid- to large-sized automated libraries in the united states and abroad use machine readable cataloging (marc) records to form the basis of their online catalogs. bibliographic records, including those in the marc format, generally represent an individually published item, or “information product,” and describe the physical characteristics of the item itself. the basic unit of archival description, however, is a much more complex entity than the basic unit of bibliographic description and often involves multiple hierarchical levels that may or may not extend down to the level of individual items. at portland state university (psu) the authors examined whether the capabilities of their present integrated library system could be expanded to capture the hierarchical structure of traditional archival finding aids. ■ background as early as , the cataloging rules established by panizzi were geared toward locating individual pub- lished items. panizzi based his rules on the idea that any person looking for any particular book should be able to find it through the catalog. this tradition has con- tinued over time up through current standards such as the anglo-american cataloguing rules and reaffirmed in marc, the standard for the representation and exchange of bibliographic information that has been widely used by libraries for over thirty years. archival description, on the other hand, is generally based on the fonds, that is, the entire collection of materi- als in any medium that were created, accumulated, and used by a particular person, family, or organization in the course of that creator’s activities and functions. thus, the basic unit of archival description, usually a finding aid, is a much more complex entity than the basic unit of biblio- graphic description, often involving multiple hierarchical levels of description that may or may not extend down to the level of individual items. before archival description begins, the archivist iden- tifies related groups of materials and determines their proper arrangement. once the arrangement is deter- mined, then the description of the materials reflects both their provenance and their original order. the first explicit statement of the levels of arrangement in an archi- val collection was by holmes and has since been elevated to the level of dogma in the archival community. a more recent statement in describing archives: a content standard (dacs) indicates that the actual levels of arrangement may differ for each collection. by custom, archivists have assigned names to some, but not all, levels of arrangement. the most commonly identified are collection, record group, series, file (or filing unit), and item. a large or complex body of mate- rial may have many more levels. the archivist must determine for practical reasons which groupings will be treated as a unit for purposes of description. rephrasing holmes, the five levels of arrangement can be defined as: . the collection level which holmes called the depos- itory level—the breakdown of the depository’s complete holdings into a few major divisions based on the broadest common denominator . the record group level—the fonds or complete col- lection of the papers of a particular administrative division or branch of an organization or of a par- ticular individual or family . the series level—the breakdown of the record group into natural series and the arrangement of each series with respect to the others . the filing unit level—the breakdown of each series into unit components, which are usually fairly obvious if the documents are kept in file folders . the document level—the level of individual items digital collection management through the library catalog michaela brenner, tom larsen, and claudia weston digital collection management through the library catalog | brenner, larsen, and weston michaela brenner (brennerm@pdx.edu) and tom larsen (larsent@pdx.edu) are database maintenance and catalog librarians, and claudia weston (westonc@pdx.edu) is assis- tant university librarian for technical services, portland state university. information technology and libraries | june the end result of archival description is usually a find- ing aid that ideally presents an accurate representation of the items in an archival collection so that users can, as independently as possible, locate them. building on the print finding aid, the archival com- munity has explored a number of mechanisms for disseminating information on the availability of items in their collections. in , the usmarc format for archival and manuscript control (marc-amc) was released and subsequently sanctioned for use as one possible standard data structure and communication protocol in the saa descriptive standard archives, personal papers, and manuscripts (appm) and its succes- sor, dacs. its adoption, however, has been somewhat controversial among archivists. the difficulty in capturing the hierarchical nature of collections through the marc format is one factor that has limited the use of marc by the archival community. while it is possible to encode this hierarchical description in marc using notes and linking fields, few archivists in practice have actually made use of these linking fields. thus, in archival cataloging, marc records have been used primarily for collection-level description, allowing users to search and discover only general information about archival collections in online catalogs while the finding aid has remained the primary tool for detailed data at all levels of description. in , the encoded archival description (ead) emerged as a new standard for encoding descriptions of archival collections. the ead standard, like the marc standard, allows for the electronic storage and exchange of archival information; but unlike marc, it is based on the finding aid. ead is well suited for encoding the hierarchical relationships between the different parts of the collection and displaying them to the user, and it has become more widely adopted by the archival com- munity. as outlined, the standards and systems chosen by an institution are dictated by the needs and traditions of that institution. the archival community relies heavily on finding aids and, with increasing frequency, on ead, their electronic extension; whereas the library commu- nity heavily relies on the online public access catalog (opac) and marc records. new trends capitalizing on the strengths of both traditions are evolving as libraries and archives seek ways to improve access to their archi- val and digital collections. ■ access to digital archival collections in libraries when searching the web for collections of informa- tion, one frequently encounters separate interfaces for traditional library, archival, and digital collections even though these collections may be owned, sponsored, hosted, or licensed by a single institution. descriptive records for traditional library materials reside in the opac and are constructed according to standard library practice, while finding aids for the archival and digital collections increasingly appear on specially designed web sites. this, of course, means that users searching the opac may miss relevant materials that are described only in the archival and digital documents database or web site. similarly, users searching the archival and digi- tal documents database or web site may miss relevant materials that are described only in the opac. in other instances, libraries, such as the library of congress, selectively add records to their opacs for indi- vidual items in their archival and digital document col- lections. this incorporation allows users more complete access to items within the library’s collections. authority control and the assignment of descriptors further enhance access to the item-level records. to minimize processing costs, however, libraries frequently create brief descrip- tive records for items, thereby limiting their value to patrons. by creating descriptive records for the items only, libraries also obscure the hierarchical relationships among the items and the collections in which they reside. these relationships can provide the user with a useful context for the individual items and are an essential part of archival description. still other libraries, such as the university of washing- ton, include collection-level marc records in the opac for their archival and digital document collections. these are searchable in the opac in the same way as biblio- graphic records for other materials. these collection-level records can then in turn be linked to finding aids that describe the collections more fully. collection-level records often are used in libraries where library resources may be insufficient for cataloging large collections of materials at the item level. the guidelines for collec- tion-level records in appm and dacs, however, allow for additional fields that are not ordinarily used in library bibliographic records. these include such things as descriptions of the organization and arrangement of the collection, citations for published descriptions of the collection and links to the finding aid, and acknowledg- ment of the donors, as well as ample subject access to the collection. despite their potential for detail, collection- level records cannot provide the same degree of access to individual items as full item-level records. ■ an approach taken at portland state university library in many ways, archival and digital-document collections are continuing resources. a continuing resource is defined as “. . . a bibliographic resource that is issued over time digital collection management through the library catalog | brenner, larsen, and weston with no predetermined conclusion. continuing resources include serials and ongoing integrating resources.” like published continuing resources, archival and digital collections generally are created over time with no predetermined conclusion. in fact, some archival col- lections continue to grow even after part of the collection has been accessioned by a library or archive. thus, even though many of the individual items in the collection might be properly treated as monographic (not unlike serial analytics), it would not be unreasonable to treat the entire collection as a continuing resource. with this in mind, the authors examined whether their electronic-resource management system could be adapted to accommodate evolving collections of digitized and born-digital material. more specifically, the present system was examined to determine whether its capabili- ties could be expanded to capture the hierarchical struc- ture found in traditional archival finding aides. the electronic resource management system in use by psu library is innovative interfaces’ electronic resource management (erm) product. according to innovative interfaces inc.’s (iii) marketing literature, “[erm] effec- tively controls subscription and licensing information for licensed resources such as e-journals, abstracting and indexing (a&i) databases, and full-text databases.” to control and provide improved access to these resources, erm stores details about purchase orders, aggregators and publishers, subscription terms, licensing conditions, breadth of holdings, internal and external contact infor- mation, and other aspects of these resources that individ- ual libraries consider relevant. for increased security and data integrity, multilevel permissions restrict viewing and editing of data to the appropriate level of staff or patron. the ability of erm to replicate the two-level hierarchi- cal relationships between aggregators or publishers and the electronic and print resources they provide was of par- ticular interest to the authors. through erm and iii’s batch record load capabilities, bibliographic and resource records can be loaded into the iii system using delimited source files such as those provided by serials solutions. resource records are the mechanisms used by iii to describe digi- tal resources at a collection, subcollection, or title level, thereby enabling the capture of descriptive information not permitted by standard bibliographic records. iii uses holdings records to document serial holdings statements. according to the marc formats for holdings data, a holdings statement is the “record of the location(s) and bibliographic units of a specific bibliographic item held at one or more locations.” iii holdings records may also contain a url for connecting to an electronic resource. in figure , for example, the resource record shows that psu library provides limited access to a number of journal titles through its springer journals online resource. as seen in figure , the display of a holdings record embedded in a bibliographic record provides more spe- cific information on the availability of a title through the library’s collection. in this particular example, the information display reveals that print volumes are avail- able for this title but that psu only has this title avail- able as a part of the springer-verlag electronic collection accessible by clicking on the hotlink. more information on the springer collection can be discovered by clicking on the about resource button to retrieve the springer journals online resource record. this example, then, represents a two-level hierarchy where the resource springer journals online is analogous to an archival collection and abdominal imaging is analogous to an archival series. adaptation of erm for library-created digital collec- tions was explored through work being done to fulfill the requirements of a grant received in by psu library. the goal of this grant was “to develop a digital library under the sponsorship of the portland state university library to serve as a central repository for the col- lection, accession, and dissemination of key planning documents and reports, maps, and other ephemeral materials that have high value for oregon citizens and for scholars around the world.” the overall collection is called the oregon sustainable community digital library (oscdl). in addition to having its own web site, it was decided to make this collection accessible through the psu library catalog so that patrons could find digitized original documents about the city of portland together with other library materials. bibliographic records would be added to the database with hyperlinks to the digitized original documents using existing staff and tools. these biblio- graphic marc records would be as complete as possible. initially, attention was focused on documents origi- nating from four different sources: ernest bonner, a for- mer portland city planner; the city of portland archives; metro (the regional government for the portland, oregon, metropolitan area); and trimet (the portland metro- politan public transportation system). along with the documents, metadata was received from various data- bases. these descriptions ranged from almost nothing to detailed archival descriptions. unlike the challenge of shifting titles and holdings with typical serials collections, the challenge of this project was to reflect the four hierarchical levels of psu library’s collection (figure ). innovative’s system struc- ture was manipulated in order to accomplish this. at the core of iii’s erm module are resource records (rr) created to reflect the peculiarities of a particular collection. linked to these resource records are holdings records (hr) containing hyperlinks to the actual digi- tized documents (doc h – doc h ) as well as to their respective bibliographic records (bib doc h – bib doc h ) containing additional information on the individual items within the collection (figure ). information technology and libraries | june first, resource records were manually created for three of the subcollections within the bonner collection. these subcollections contained documents reflecting the development of harbor drive, front street, and the park blocks. the fields defined for the resource records include the resource title; type (digitized documents) and format (pdf) of the resource; a hyperlink to the new oscdl web site; content and systems contact names; a brief descrip- tion of the resource; and, most importantly, the resource id used to connect holding records for individual docu- ments to the corresponding resource record. next, the batch-loading function in erm was used to create bibliographic and holding records and associ- ate them with the resource records. taking advantage of tracking data produced during the digitization process (figure ), spreadsheets were created for each collection reflecting the data assigned to each individual digitized document. the document title, the date the document was created, number of pages, and summaries were included. coordinates for the streets mentioned in the documents were also included. because erm uses issn numbers and titles as match points for record loads, ”issn” numbers were also manufactured for each docu- ment and included in the spreadsheet. these homemade numbers were distinguished by using pdx as a prefix followed by collection and document numbers or letters, for example, pdx or pdxhdcoll. fortunately, erm accepted these dummy issns (figure ). from this data spreadsheet, the system-required comma delimited coverage load file (*.csv) was also created. for this file, the system only allows a limited number of fields, and is very particular about the right terms, including correct capitalization, for the header row. individual document titles, the made-up issn numbers, individual urls to the documents, and a collection-specific resource id (provider) that connects all the documents from a collection to their respective resource record were included. the resource id is the same for all documents in one collection (figure ). in the first attempt, the system was set up to produce holdings and bibliographic records automatically, using the data from the spreadsheets. for the bibliographic records, a system-provided template was created that included some general subject headings, genre headings, an author field, and selected fixed fields, such as language, bibliographic level, and material type (figure ). records for the harbor drive collection were loaded, and the system created brief bibliographic and holdings records and linked them to the harbor drive resource record. the records were globally updated to add the general material designator (gmd) “electronic resource” to the title as well as the phrase “digitized document” as a local “call number” to make these documents more visible in the browse screen of the online catalog (opac) (figure ). the digitized documents now could be found in the library catalog by author, subject, or keyword. the brief bibliographic records (figure ) allow the user to go either to the digitized document via url or to the resource record with more information on the resource itself and links to other items in the same collection. the resource record then provides links either to the new oscdl web site (via the - oregon sustainable community digital library link at the bottom of the resource record), to the bibliographic description of the individual document, or to the digitized document (figure ). however, the quality of the brief bibliographic re- cords that had been batch generated through the sys- tem-provided template was not satisfactory (figure ). it was decided that more document-specific data like summaries, number of pages, the dates the documents were created, geographical information, and document- level local subject headings should be included. these data were already available from the original spread- sheets. with limited time and staff resources, full bib- liographic marc records were batch created using the spreadsheets, detailed templates adjusted slightly to each collection, microsoft mail merge, and finally, the marcedit program created by terry reese of oregon state university (http://oregonstate.edu/~reeset/mar- cedit/html/index.html). this gave maximum control over the data to be included and the way they would be included. it also eliminated the need to clean up the data following the record load (figure ). subsequently, full bibliographic records were created for the subcollections harbor drive, front street, and park blocks, to connect them to the next higher level, the bonner collection (figure ). these records were also contributed to worldcat. mimicking the process used at the document level, a resource record was created for the bonner collection and the holdings records for the three subcollections were connected with their corresponding bibliographic records (figure ). resource records with their corresponding item-level records for trimet, the city archives, and metro fol- lowed. the final step was then to add the resource record and the bibliographic record for the whole oscdl col- lection (figure ). since this last bibliographic record is not connected to a collection above it, there is only a hyperlink to the oscdl resource record (figure ). more subcollections and their corresponding digi- tal documents are continually being added to oscdl. structures in psu library’s opac are adjusted as these collections change. digital collection management through the library catalog | brenner, larsen, and weston ■ conclusion according to salter, “digitizing, the current challenge that straddles the th and st centuries, has given archi- vists and librarians pause to reconsider access to their collections. the world of digitization is the catalyst for it people, librarians, and archivists to unify the way they do things.” in this paper, a strategy has been offered for adapting a library system to traditional archival practice. by making use of some of the capabilities of the module in psu library’s integrated library system that was originally designed for managing electronic resources, a method was developed for managing digital archival col- lections in a way that incorporates some of the features of a traditional finding aid. the contents of the various hierarchical levels of the collection are fully represented through the manipulation of the record structures avail- able through psu’s system. this technique provides for enhanced access to the individual items of a collection by giving the context of the item within the collection. links between the hierarchical levels facilitate navigation between the levels. although the records created for traditional library systems are not as rich as those found in traditional finding aids, or in ead, their electronic equivalent; and the visual arrangements are not as intriguing as a well- planned web site, the ability to show how items fit within the greater context of their respective collection(s) is a step toward reconciling traditional library and archival practices. enabling the library user to virtually browse through the overall resources offered by the library and then, if desired, through the various levels of a collection for relevant resources enhances the opportunities pre- sented to the user for finding relevant information. references and notes . society of american archivists, “so you want to be an archivist: an overview of the archival profession,” , www.archivists.org/prof-education/arprof.asp (accessed apr. , ). . kent m. haworth, “archival description: content and context in search of structure,” journal of internet cataloging , no. / ( ): – . . antonio panizzi, “rules for the compilation of the cata- logue,” the catalogue of the british museum ( ): v–ix. . joint steering committee for revision of aacr, anglo- american cataloguing rules, nd ed., revision (chicago: ala, ). . society of american archivists, describing archives: a content standard (chicago: society of american archivists, ). . haworth, “archival description.” . oliver w. holmes, “archival arrangement: five different operations at five different levels,” american archivist , no. ( ): – ; terry abraham, “oliver w. holmes revisited: levels of arrangement and description of practice,” american archivist , no. ( ): – . . society of american archivists, describing archives: a content standard (chicago: society of american archivists, ); xiii. . haworth, “archival description.” . society of american archivists, describing archives: a content standard (chicago: society of american archivists, ); steven l. hensen, comp., archives, personal papers, and manuscripts, nd ed. (chicago: society of american archivists, ). . peter carini and kelcy shepherd, “the marc standard and encoded archival description,” library hi tech , no. ( ): – ; steven l. hensen, “archival cataloging and the internet: the implications and impact of ead,” journal of inter- net cataloging , no. / ( ): – . . abraham, “oliver w. holmes revisited.” . elizabeth j. weisbrod and paula duffy, “keeping your online catalog from degenerating into a finding aid: con- siderations for loading microformat records into the online catalog,” technical services quarterly , no. ( ): – . . carini and shepherd, “the marc standard and encoded archival description.” . see, for example, margaret f. nichols, “finding the forest among the trees: the potential of collection-level catalog- ing,” cataloging & classification quarterly , no. ( ): – ; and weisbrod and duffy, “keeping your online catalog from degenerating into a finding aid.” . joint steering committee for revision of aacr, anglo- american cataloguing rules, d- . . innovative interfaces inc., “electronic resources manage- ment,” , www.iii.com/pdf/lit/eng_erm.pdf (accessed apr. , ). . library of congress, marc format for holdings data: including guidelines for content designation (washington, d.c.: cataloging distribution service, library of congres, ), appendix e–glossary. . carl abbot, “planning a sustainable portland: a digital library for local, regional, and state planning and policy documents—framing paper,” , http://oscdl.research.pdx. edu/framing.php (accessed apr. , ). . anne a. salter, “ st-century archivist,” newsletter, , www.lisjobs.com/newsletter/archives/sept asalter.htm (accessed apr. , ). information technology and libraries | june figure . example of resource record from the psu library catalog (search conducted nov. , ) appendix. figures digital collection management through the library catalog | brenner, larsen, and weston figure . example of a bibliographic record for a journal title from the psu library catalog (search conducted nov. , ) information technology and libraries | june figure . resource record harbor drive with linked holdings records, bibliographic records, and original documents figure . partial diagram of the hierarchical levels of the collection digital collection management through the library catalog | brenner, larsen, and weston figure . comma delimited coverage load file (*.csv) figure . data spreadsheet figure . spreadsheet for tracking data information technology and libraries | june figure . browse screen in opac figure . bibliographic records template digital collection management through the library catalog | brenner, larsen, and weston figure . resource record with various links figure . system-created brief bibliographic record in opac information technology and libraries | june figure . bonner resource record with linked holdings records, bibliographic records, and original documents figure . full bibliographic record in opac digital collection management through the library catalog | brenner, larsen, and weston figure . bibliographic record for the oscdl collection figure . outline of linked records in the collection information technology and libraries | june in the early years of modern information retrieval, the fundamental way in which we understood and evalu- ated search performance was by measuring precision and recall. in recent decades, however, models of evaluation have expanded to incorporate the information-seeking task and the quality of its outcome, as well as the value of the information to the user. we have developed a systems engineering-based methodology for improving the whole search experience. the approach focuses on understand- ing users’ information-seeking problems, understand- ing who has the problems, and applying solutions that address these problems. this information is gathered through ongoing analysis of site-usage reports, satisfac- tion surveys, help desk reports, and a working relation- ship with the business owners. ■ evaluation models in the early years of modern information retrieval, the fundamental way in which we understood and evalu- ated search performance was by measuring precision and recall. in recent decades, however, models of evaluation have expanded to incorporate the information-seeking task and the quality of its outcome, cognitive models of information behavior, as well as the value of the informa- tion to the user. the conceptual framework for holistic evaluation of libraries described by nicholson defines multiple perspectives (internal and external views of the library system as well as internal and external views of its use) from which to measure and evaluate a library system. the work described in this paper is consistent with these frameworks as it emphasizes that, while efforts to improve search may focus on optimizing preci- sion or recall, it is equally important to recognize that the search experience involves more than a perfect set of high-precision, high-recall search results. the total search experience and how well the system actually helps the user solve the search task must be evaluated. a search experience begins when users enter words in a search box. it continues when the users view some representation (such as a list or a table) of candidate answers to their queries. it includes the users’ reactions to the usefulness of those answers and their representa- tion in satisfying information needs, and continues with the users clicking on a link (or links) to view content. optimizing search results without considering the rest of the search experience and without considering user behavior is missing an opportunity to further improve user success. for example, the experience is a failure if typical users cannot recognize the answers to their infor- mation need because the items lack a recognizable title or an informative description, or they involve extensive scrolling or hard-to-use content. ■ proposed solutions problems with search, such as low precision or low recall, are often addressed by either metadata solutions (add- ing topical tags to content objects based on controlled vocabularies) or replacement of the search engine. the problems with the metadata approach include the time and effort required to establish, evolve, and maintain taxonomies, and the need for trained intermediaries to apply the tags. a community of stakeholders may be convened to define the controlled vocabulary, but often the lowest common denominator prevails, the champi- ons and stakeholders leave, and no one is happy with the resulting standard. even with trained intermediaries, inter-indexer inconsistency compromises this approach, and inconsistent term application can cause degradation of search results. another shortcoming of the metadata approach is that a specific metadata classification is just a snapshot in time and assumes that there is only one particular hierarchy of the information in the corpus. in reality, however, there is almost always more than one way to describe a concept, and the taxonomy is the view of only one individual or group of individuals. in addition, topical metadata is often implemented with little understanding of the types of queries that are submitted or the probable user search behavior. the other approach to improving search results— replacing a search engine—is not a guarantee to fixing the problem because it focuses only on improving precision (and perhaps recall as well) without understanding the true barriers to a successful search experience. ■ irs.gov irs.gov, one of the most widely used government web sites, is routinely accessed by millions of people each month (more than million visits in april ). as an informational site, the key goal of irs.gov is to direct visitors quickly to useful information, either through marcia d. kerchner (mkerchner@mitre.org) is a principal information systems engineer at the mitre corporation, mclean, va. a dynamic methodology for improving the search experience marcia d. kerchner article title | author a dynamic methodology for improving the search experience | kerchner navigation or a search function. given that there were almost million queries submitted to irs.gov in april , search is clearly a popular way for its users to look for information. this paper offers an alternative to con- ventional search-improvement approaches by presenting a systems engineering-based methodology for improv- ing the whole search experience. this methodology was developed, honed, and modified in conjunction with work performed on the irs.gov web site over a three- year period. a similar strategy of “sense-and-respond” for information technology (it) departments of public organizations that involves systematic intelligence gath- ering on potential customer demand, a rapid response to fulfill that demand, and metrics to determine how well the demand was satisfied, has recently been described. the methodology described in this paper focuses on analyzing the information-seeking behaviors and needs of users and determining the requirements of the busi- ness owners (the irs business operating divisions that provide content to irs.gov, such as small business and self-employed, wage and investment) for directing users to relevant content. it is based on the assumption that a web site must evolve based on its user needs, rather than expecting users to adapt to its singularities. to sup- port this evolution, this approach leverages techniques for query expansion and document-space modification. dramatic improvements in quality of service to the user have resulted, enhancing the user experience at the site and reducing the need to contact the help desk. the approach is particularly applicable for those government, corporate, and commercial web sites where there is some control over the content, and usage can be categorized into regular patterns. the rest of this paper provides a case study in the application of the methodology and the application of metrics, in addition to precision and recall, to measure search experience improvement. ■ conceptual framework while analysis of search results often focuses on search syntax and search-engine performance, there are actu- ally several steps in the retrieval process, from the user identifying an information need to the user receiving and reviewing query results. as shown in figure , find- ing information is a holistic process. there are several opportunities to improve the whole user experience by fine-tuning this process with a variety of tools—from document engineering to results categorization. once the user and business-owner needs are understood, the appropriate tools to address specific issues can be identified. the tools in our toolkit are described in the follow- ing sections. document engineering document engineering includes: ■ document-space modification: modifying the docu- ment space by adding terms to content (especially to titles) that are good discriminators and reflect terms commonly entered by users. this approach has the added benefit of making the content more under- standable to users. ■ establishment of content-quality standards: defining business processes that improve content quality and organization. document-space modification there is significant syntactic and semantic imprecise- ness in the english language. in addition, because of the inadequacies of human or automatic keyword assign- ment, standard means of representing documents in indexes by statistical term associations and frequency counts or by adding metadata tags are not definitive enough to produce a space that is an exact image of the original documents. document-space modifica- tion moves documents in the document space closer to future similar queries by adding new terms or modify- ing the weight of existing terms in the content (figure ). the document space is thus modified to improve retrieval. for irs.gov, rather than adjusting content weights, titles and content are modified to adjust to changing terminology and user needs. establishment of content-quality standards the quality of the search correlates with the quality of the content. improved search results can be achieved by applying good content-creation practices. retrieval can be significantly improved by addressing problems observed in the content. these problems include inconsistencies in term use—for example, earned income credit (eic) ver- sus earned income tax credit (eitc)—duplicate content, insufficiently descriptive page titles, missing document summaries, misspellings, and inconsistent spellings. figure . the information retrieval process information technology and libraries | june processes to improve content quality should estab- lish standards for consistent term usage in content, as well as standards for consistent and descriptive naming of content types (for example, irs types include forms, instructions, and publications). these processes will not only improve search precision, but will also help users identify appropriate content in the search results. for example, content entitled “publication ” in response to the query “child care” may be the perfect answer (with excellent precision and recall), but the user will not recog- nize it as the right answer. a title such as “publication : child and dependent care expenses” will clearly point the user to the relevant information. usability tests conducted in march for irs.gov confirmed that content organization plays an important role in the perceived success of a user’s search experi- ence. long pages of links or scrolling pages of content left some users confused and overwhelmed, unable to find the needed information. for these queries, although the search results were perfect, with a precision of percent after one document, the search experiences were still failures. query enhancement the technique of relevance feedback for query expansion improves retrieval in an iterative fashion. according to this approach, the user submits a query, reviews the search results, and then reports query-document rel- evance assessments to the system. these assessments are used to modify the initial query, that is, new terms are added to the initial query (hopefully) to improve it, and the query is resubmitted. if one visualizes the content in a collection as a space (figure ), this approach attempts to move the query closer to the most relevant content. a drawback of relevance feedback is that it is not generally collected over multiple user sessions and over time, so the next user submitting the same query has to go through the same process of providing results evalu- ations for query expansion. borlund has noted that, given that an individual user ’s information need is personal and may change over session time, relevance assessments can only be made by a user at a particular time. however, on irs.gov, where there are many common queries for which there is a clear best-guess response, there is valuable relevance information that, if captured once, could benefit tens of thousands of users for specific queries. in fact, in april , the top four hundred queries represented almost half of all the queries. another drawback of the relevance-feedback ap- proach is that it forces the user, novice or expert, to become engaged in the search process. as noted previ- ously, users are generally not interested in becoming search experts or in becoming intimately involved in the process of search. the relevance-feedback approach tries to change users’ behavior and forces them to find the specific word or words that will best retrieve the relevant information. in fact, some research has shown that the potential benefits of relevance feedback may be hard to achieve primarily because searchers have difficulty find- ing useful terms for effective query expansion. to avoid requiring users to submit relevance-feedback judgments, the methodology uses alternative approaches for gathering feedback: ( ) mining sources of input that do not require any additional involvement on the part of the users; and ( ) soliciting relevance judgments from subject matter experts. as noted above, while best results may be different per task and per user, particularly given the shortness of the queries, our goal is to maximize the good results for the maximum number of people. best-guess results are derived from a variety of sources, including usability testing, satisfaction survey questionnaires, and business- content owners. for example, users entering the common query “ ez” can be looking for information on the form or the form itself. given that—as shown in table (based on the responses of , users to satisfaction sur- veys in )—the goal of percent of irs.gov searchers is to download a form as opposed to percent seeking to obtain general tax information, the retrieval of the ez form and its instructions is prioritized, while also retrieving any general related information. figure . document-space modification figure . query modification article title | author a dynamic methodology for improving the search experience | kerchner we can determine the best-guess results as follows: ■ review the search results for terms that are on the frequently entered search-terms list ■ review help desk contacts, satisfaction-survey com- ments, and zero-results reports to identify infor- mation users who are having trouble finding or understanding ■ identify best results by working with the business owners as necessary ■ analyze why best results are not being retrieved for a particular query ■ add appropriate synonyms for this and relat- ed queries ■ engineer relevant documents (as described above) in this way, the thesaurus, as the source for query enhancement, is an evolving structure that adapts to the needs of the users rather than being a fixed entity of elements based on someone’s idea of a standardized vocabulary. search improvement we can intercept very popular queries and return a set of preconfigured results or a quick link at the top of the search-results listing. for example, the user enter- ing “ ” sees a list of the most popular -related forms and instructions in addition to a list of other search results. there were more than , users in april who requested the i- form. since the form is not an irs form, users are presented with a link to the bureau of citizen and immigration services web site. the tens of thousands of users who look for state tax forms on irs.gov are directed either to the spe- cific state-tax-form web- site page or to a page with links to state tax sites. this unique and user-friendly approach provides a sig- nificant improvement over a page that tells the user that there is no matching result, leaving him to fend for himself. another technique for improving search preci- sion (not currently used for irs.gov) is to tune and adjust parameters in the search engine, such as the relative weighting of basic metadata tags such as title (if they are used in the relevance calculation). results-ranking improvement the search results can be programmatically re-ranked before being presented to the user. this approach (not used as yet on irs.gov) is a variation on the quick links described above for re-ranking more than one result. categorization a large set of search results can be automatically catego- rized into subsets to help the user find the information he needs. in addition, a “search within a search” function is available to help the user narrow down results. research to be conducted on commercial products to support auto- matic categorization is planned for the future. summarization as noted earlier, a barrier to a successful user experience can be the lack of informative descriptions in the search results. therefore, an important tool for search-experi- ence improvement is to make sure that content titles and summaries are informative, or as a second choice, that the search engine dynamically generates informative sum- maries. passage-based summaries and highlighted search terms in the summary and the content have become a feature of many commercial search engines as another way to improve the usability of the returned results. in table . reasons for using irs.gov reason for coming to irs.gov % of total site visitors % of total search users download a tax form, publication, or instructions obtain general tax information obtain information on e-file other obtain info on tax regulations or written determinations order forms from the irs sign up or login to e-services link and learn (vita/vce) training obtain info on the status of your tax return use online tax calculators obtain info on revenue rulings or court cases obtain an employer identification number (ein) — note: due to rounding, totals may not equal %. information technology and libraries | june addition, for those pdf publications that lacked informa- tive titles in the title tag, descriptive information from a different metadata field was added to the search display programmatically, which improved the usability of such results significantly. ■ methodology the methodology for evolving the search functionality is based on a logical, systems-engineering approach to the issue of getting users the information they seek: under- standing the problems, understanding who has the prob- lems, and applying solutions that address the problems. usability studies, weblogs, focus groups, help desk contacts, and user surveys provide different perspectives of the information system. the steps of the methodology are: . understand the user population. . identify the barriers to a successful search experi- ence. . analyze the information-seeking behaviors of the users. . understand the needs of the business owners. . identify and use the appropriate tools to improve the user’s search experience. . repeat as needed. . monitor new developments in search and analytic technologies and replace the search engine as appropriate. step : understand the user population the first step is to profile and understand the user popu- lation. as mentioned above, an online satisfaction survey was conducted during a six-week period in january– february , to which , users responded. the users were asked the frequency of their usage of the site, their primary reason for coming to irs.gov, their category (individual, tax professional, business representative), and how they generally find information on irs.gov. as shown in tables – , percent of the irs. gov visitors use it once a month or less (the largest group being those who use it every six months or less), or were using it for the first time; percent are individual taxpayers; percent are tax profes- sionals; percent visit the site to download a form or publication; and percent come for general tax or e-file information. forty-nine per- cent use the search engine. not surprisingly, percent of the frequent visitors (those who visit once a week or more) are tax professionals, while percent of the infrequent visitors are individuals or those who represent a business. the most common task of both the most frequent and infrequent visitors is to download a form, publication, or instructions, followed by obtaining general tax informa- tion. most frequent and infrequent visitors use the search function to locate their information. thus, the largest group of irs.gov users consists of average citizens, unfamiliar with the site, who have a specific question or a need for a specific form or publication. these users require high-precision, highly relevant results, and a highly intuitive search interface. they do not want or need to read all the material gen- erated by their search, but they want their question answered quickly. these users are generally not experienced with so- phisticated query language syntax, and because they come to the site no more than once a month, they are not likely to be familiar with its navigational organization. as studies demonstrate, users in general do not want to learn a search engine interface or tailor their queries to the design of a particular search engine. they want to find their information now before “search rage” sets in. one study observed that, on average, searchers get frus- trated in twelve minutes. tax professionals form a small but important group of irs.gov users that includes lawyers, accountants, and tax preparers. they generally use the site on a regular basis, which could be daily, weekly, or monthly. some of these users, particularly lawyers and accountants, require high relevance in their search results; it is critical that they retrieve every relevant piece of information (e.g., all the tax regulations) related to a tax topic. they may be willing to sift through large results sets to make sure they have seen all the relevant items. in contrast, many tax preparers use the site primarily to download forms and instructions. while these different sets of users have different levels of expertise using the site and somewhat different precision and recall requirements, they do have one char- acteristic in common—they are not interested in search table . frequency of visits to irs.gov first time every six months or less about once a month about once a week daily more than once a day site visitor % % % % % % search user % % % % % % article title | author a dynamic methodology for improving the search experience | kerchner for its own sake. approaches to improving retrieval results that focus on forcing users to use tools to refine their query to get presumably better search results (e.g., leveraging the power of boolean or other search syntax) are not desirable in a public web site environment. the complexity of the search must be hidden behind the search box and users must be helped to find information rather than be expected to master a search function. step : identify the barriers to a successful search experience there are several categories of reasons why finding infor- mation on a public web site can be frustrating for the user. ■ mismatch between user terminology and content terminology  the user search terms may not match the ter- minology or jargon used in the content (e.g., users ask for “tax tables” or “tax brackets”; the irs names them “tax rate schedules”).  multiple synonymous terms or acronyms are found because different authors are pro- viding content on similar topics (e.g., “ein,” “employer identification number,” “federal id number”; “eic” versus “eitc”).  users request the same information in a vari- ety of ways (e.g., “ ez,” “ -ez,” “ez,” “form ez,” “ ez form,” “ ez,” “ez ”).  related content may be inconsistently named, complicating the user’s search process (e.g., “ x” form versus “ -x” instructions).  the user may use a familiar acronym that is spelled out in the content (e.g., “poa” for “power of attorney”). ■ mismatch between user requests and actual content  many users ask for information that they ex- pect to find on the site but is actually hosted at another site (e.g., “ds ,” a department of state form; “it- ,” a new york state tax form). ■ issues with results listing and content display  content may lack informative titles.  automatically generated summaries may not be sufficiently descriptive for users to recog- nize the relevant material in the results listing.  content may consist of long, scrolling pages, which users find hard to manage. ■ incomplete user queries  very short search phrases (average length of less than two words) can make it difficult for a search algorithm to deduce the specific con- tent the user is seeking. step : analyze the information-seeking behaviors of the users site-usage reports, satisfaction surveys, help desk con- tact reports, zero-results reports, focus groups, and usability studies are valuable sources of information. they should be mined for information-seeking behaviors of the site’s users and other barriers to a successful search experience, as follows: ■ review site-usage reports for the most frequently entered search terms and popular pages (both may change over time) and the zero-results search terms. look for:  new terms  variations on popular terms  common misspellings or typos  common searches, including searches for items table . irs.gov user types type of user % of total site visitors % of total search users individual taxpayer % % representing a business % % tax professional % % representing a charity or nonprofit % % vita/vce volunteers % % representing a government entity % % student % % irs employee % % other % % table . how users find information on irs.gov how do you usually find information on irs.gov? % of total site visitors search engine % irs keyword % navigation to the web page % internet search engine (e.g., google, yahoo) % site map % other % bookmarks % links to irs.gov from other web sites % information technology and libraries | june not on the site, that could be candidates for pre- programmed “quick links”  frequently entered terms—review search re- sults to identify candidates for improvement ■ review satisfaction surveys over time  look for new problems that caused satisfac- tion to decrease  analyze answers to questions asking what people could not find, potentially identifying new barriers to success ■ conduct usability studies  identify issues with the user interface as well as with content findability and usability ■ review help desk contact reports  identify which topics users are having trouble finding or understanding step : understand the needs of the business owners the business owners are the irs business operating divisions that provide content to irs.gov, such as small business and self-employed, wage and investment. it is important to involve them in the process of enhancing the user experience, because they may have specific goals for prioritizing information on a particular topic or may be managing campaigns for highlighting new informa- tion. thus it is desirable to: ■ meet with business owners regularly to understand their goals for providing information to users ■ work with them to increase the findability of their content for example, when an issue in finding a particular content topic is identified (e.g., through an increase in help desk contacts), one approach is to show the business owner the actual results that common queries (based on the site-usage reports) on the topic retrieve and then pres- ent suggested alternative results that could be retrieved with a variety of enhancement techniques, such as thesau- rus expansion or title improvement. the business owner can then evaluate which set of results presents the content in the most informative manner to the user. steps – facilitate work behind the scenes to gather the data needed to improve precision and recall and to make information more findable. the remaining steps use these data to adapt proven, widely used techniques for improv- ing search experiences to a web site’s specific environment. step : identify appropriate tools to improve the information-retrieval process as described in the previous section, the tools in our toolkit are document engineering, query enhancement, search improvement, results-ranking improvement, cat- egorization, and summarization. step : repeat as needed the process of improving the user search experience is ongoing as the site evolves. at irs.gov, different search terms appear on the site-usage reports over time, depend- ing on whether or not it is filing season, or as new content and applications are published. human intervention (with the help of applicable tracking software) is essen- tial for incorporating business requirements, evaluating human behavior, and identifying changing terms. step : monitor new developments in search and analytic technologies and replace the search engine as appropriate although a new search engine will not address all the issues that have been described, new features such as passage-based summaries and term-highlighting can improve the search experience. of course, one should consider replacing a search engine if new technology can demonstrate significantly improved precision and recall. the application of the methodology and the use of the toolkit for irs.gov will be described in the next section. ■ findings site-usage reports in , an example of a serious mismatch in user and content terminology was discovered when site-usage reports were analyzed. users entering the equivalent terms ein, employer number, employer id number, and employer identification number retrieved significantly different sets of results. we met with the business owner, who identified a key-starting page that should be retrieved along with other highly relevant pages for all of these query terms. we recommended that “ein” be added to the title of the key page because, although ein is a very popular query, the acronym was not used in the content, but was instead spelled out. as a result, the key page was not being retrieved. synonyms were added to the query enhancement thesaurus to accommodate the variants on the ein concept. after these steps were implemented, the results were as follows: ■ for the query ein, the target page moved from # to # ■ for the query ein number, it moved from # to # article title | author a dynamic methodology for improving the search experience | kerchner ■ for the query employer identification number, it moved up to # (it was not in the top previously) ■ all search results now retrieved on the first page for these terms were highly relevant in january , there were approximately twenty thousand queries using these terms, so the search experi- ence has been improved for tens of thousands of users in one month and hundreds of thousands of users through- out the year. ■ review of help desk contacts help desk reports summarize, for each call or e-mail, the general topic of the user’s contact (filing information, employer id number, forms, and publications issues) and the specific question. for example, the report might indicate that a user needed help in finding or download- ing the w- form or did not understand the instructions for amending a tax return. as help desk contact reports were reviewed, clusters of questions emerged indicating information that many users could not find or under- stand. by analyzing approximately , contacts (e-mail, telephone, chat) during a peak five-day period in april , four particular areas were identified that were ripe for improvement: users could not find previous years’ forms, which, although they can be found on the site, are not indexed and thus not findable through search; users had questions about where to send their tax returns; users had questions about getting a copy of their tax return or w- form; and users had problems finding the x or ez forms. utilizing the information retrieval toolkit, the follow- ing improvements were implemented: a) search for previous years’ forms tool used: results-ranking improvement a user requesting a previous year’s forms (for exam- ple, misc) is now presented with a link directly to the page of forms for that specific year, as follows: recommendation(s) for: misc ■ forms and publications forms, instructions, and publications available in pdf format b) request for filing address tools used: document engineering and query en- hancement a new “where to file” page was created. synonyms were added to the thesaurus to accommodate the varia- tions on how people make this request (address, where to send, where to mail) and to prioritize retrieval of the “where to file” page. c) request for information about obtaining a copy of a tax return or w- form tools used: results-ranking improvement and query enhancement a “quick link” was created to the target page for get- ting a copy of returns and w- forms and synonyms were added to the thesaurus to prioritize related content for any query containing the word “copy.” d) requests for x or ez forms or instructions tool used: query enhancement synonyms were added to the thesaurus to address both the variations on how users requested the x and ez forms and instructions, and the inconsisten- cies in the titling of these documents (for example, the form and the instructions have different variations of the compound name). ■ results in , approximately , contacts were reviewed with the help desk during the same time period (the week before april ) to see whether the changes actually did help users find the information. it should be noted that, during this period from april to april , many other improvements to the user search experience based on the methodology were deployed. although the number of visits to irs.gov increased by approximately percent compared with the same period in , the total number of contacts with the help desk decreased by percent (there were approximately , contacts in this period in ). the results for the specific improvements are shown in table . the average decrease in contacts for those four topics was percent, compared with the average decrease of percent. this approach has significantly improved the user experience by identifying and addressing subject areas users have trouble finding or understanding on irs.gov, eliminating the need for them to contact the help desk. as a result, an increase of resources at the help desk was avoided and, hopefully, user satisfaction improved. information technology and libraries | june ■ conclusions while the case presented in this article was specific to irs.gov, the methodol- ogy itself has wide appli- cation across domains. customer service for most government and commercial organizations depends on providing users with relevant infor- mation effectively and efficiently. there are many aspects to achieving this elusive goal of matching users with the specific information they need. in this paper, it has been demonstrated that, rather than focusing just on optimizing the search engine or developing a metadata-based solution, it is essential to view the user search experience from the time content is created to the moment when users have truly found the answer to their information needs. there is no one surefire solution, and one should not assume that enhanced metadata or a new search engine is the only solution to retrieval problems. the methodology described in this paper assumes that users, especially infrequent users of public web sites, do not wish to become search experts; that intuitive interfaces and meaningful results displays contribute to a successful user experience; and that keeping business owners involved is important. the methodology is based on understanding the behavior of a site’s users in order to identify barriers to a successful search experience, and on understand- ing the needs of business owners. the methodology focuses on adapting the site to its users (rather than vice versa) through document modification, improved content-development processes, query enhancement, and targeted search improvement. it includes improve- ments to the results phase of the search process, such as improved titles and summaries, as well as to the search- and-retrieval phase. this toolkit-based approach is effective and low-cost. it has been used over the past four years to improve the user search experience significantly for the millions of irs.gov users. interesting follow-on research could focus on identifying to what degree this methodology can be automated and how to leverage new tools to pro- vide automated support for usage log analysis (such as mondosearch by mondosoft). it is clear from this case study that it is time to apply systems engineering rigor to search-experience improve- ment. this approach confirms the need to extend metrics for evaluating search beyond precision and recall to include the totality of the search experience. ■ future work teleporting has been defined as an approach in which users try to jump directly to their information targets. trying to achieve perfect search results supports the infor- mation-seeking strategy of teleporting. but the search process may involve more than a single search. people often conduct “a series of interconnected but diverse searches on a single, problem-based theme, rather than one extended search session per task.” this approach is similar to the sport of orienteering with searchers using data from their present situation to determine where to go next—that is, looking for an overview first and then submitting more detailed searches. given the general, nonspecific nature of the short queries submitted by irs.gov users, the orienteering approach may well describe the information-seeking behaviors of many users. this paper is limited to the improvement of search results for individual searches, but the need to investigate improving the search experience to support orienteering behavior is acknowledged. future research will investi- gate how to leverage the theoretical models of the infor- mation-search process, such as the anomalous states of knowledge (ask) underlying information needs and the information search process model. references and notes . “common evaluation measures,” the thirteenth text retrieval conference, nist special publication sp - (gaith- ersburg, va.: national institute of standards and technology, ), appendix a. . kalervo jarvelin and peter ingwersen, “information-seek- ing research needs extension towards tasks and technol- ogy,” information research , no. ( ), http://informationr .net/ir/ - /paper .html (accessed feb. , ); k. fisher, s. erdelez, and l. mckechnie, eds., theories of information behavior (medford, n.j.: information today, ); t. saracevic and paul b. kantor, “studying the value of library and information services, part i: establishing a theoretical framework,” journal of the ameri- can society for information science. , no. ( ): – . table . comparison of and help desk contacts problem area number of contacts number of contacts change x, ez - % prior year forms - % copy of return - % where to file - % total - % article title | author a dynamic methodology for improving the search experience | kerchner . scott nicholson, “a conceptual framework for the holistic measurement and cumulative evaluation of library services,” journal of documentation , no. ( ): – . . avra michelson and michael olson, “dynamically enabling search and discovery tem,” internal mitre presen- tation, mclean, va., mar. , . . lawrence e. leonard, “inter-indexer consistency studies, – : a review of the literature and summary of study results,” occasional paper series, no. , graduate school of library science, university of illionois, urbana-champaign, ; tefko saracevic, “individual differences in organizing, searching and retrieving information,” in proceedings of ameri- can society for information science ’ (new york: john wiley, ), – ; g. furnas et al., ”the vocabulary problem in human-system communication,” communications of the acm , no. ( ): – . . rajiv ramnath and david landsbergen, “it-enabled sense-and-respond strategies in complex public organiza- tions,” communications of the acm , no. ( ): – . . t. l. brauen et al., “document indexing based on rel- evance feedback,” report no. isr- to the national science foundation, section xi, department of computer science, cor- nell university, ithaca, n.y., ; m. c. davis, m. d. linsky, and m. v. zelkowitz, “a relevance feedback system employing a dynamically evolving document space,” report no. isr- to the national science foundation, section x, department of computer science, cornell university, ithaca, n.y., ; marcia d. kerchner, dynamic document processing in clustered collec- tions, report no. isr- to the national science foundation, ph.d. thesis, department of computer science, cornell univer- sity, ithaca, n.y., . . ibid. . gerard s. salton, dynamic information and library process- ing (englewood cliffs, n.j.: prentice-hall, ). . p. borlund, “the iir evaluation model: a framework for evaluation of interactive information retrieval systems,” information research , no. ( ), http://informationr.net/ir/ - /paper .html (accessed feb. , ). . ian ruthven, “re-examining the effectiveness of interac- tive query expansion,” in proceedings of the th international acm sigir conference on research and development in information retrieval (new york: acm press, ), – . . marc l. resnick and rebecca lergier, “things you might not know about how real people search,” , www.search tools.com/analysis/how-people-search.html (accessed oct. , ). . danny sullivan, “webtop search rage study,” the search engine report, , http://searchenginewatch.com/sereport/ article.php/ (accessed sept. , ). . j. teevan et al., “the perfect search engine is not enough: a study of orienteering behavior in directed search,” in pro- ceedings of computer-human interaction conference ’ (new york: acm press, ), – . . vicki o’day and robin jeffries, “orienteering in an infor- mation landscape: how information seekers get from here to there,” in proceedings interchi ’ (new york; acm press, ), . . n. j. belkin, r. n. oddy, and h. m. brooks, “ask for information retrieval, part i. background and theory,” the journal of documentation , no. ( ): – ; n. j. belkin, r. n. oddy, and h. m. brooks, “ask for information retrieval, part ii. results of a design study,” the journal of documentation , no. ( ): – ; carol c. kuhlthau, seeking meaning: a process approach (norwood, n.j.: ablex, ). information technology and libraries | june article title: subtitle in same font author name and second author the concept of digital libraries is familiar to both librar- ians and library patrons today. these new libraries have broken the limits of space and distance by delivering information in various formats via the internet. since most digital libraries contain a colossal amount of infor- mation, it is critical to design more user-friendly inter- faces to explore, understand, and manage their content. one important technique for designing such interfaces is information visualization. although computer-aided information visualization is a relatively new research area, numerous visualization applications already exist in various fields today. furthermore, many library professionals are also starting to realize that combining information visualization techniques and current library technologies, such as digital libraries, can help library users find information more effectively and efficiently. this article first discusses three major tasks that most visualization for digital libraries emphasize, and then introduces several current applications of information visualization for digital libraries. a good understanding of user tasks is the founda- tion of designing useful visualizations. rao et al. defined several specific user tasks of digital librar- ies and illustrated some existing information-visualiza- tion techniques that can be used to enhance these tasks, such as tilebar, cone tree, and document lens. the tasks were browsing subsets of sources iteratively, view- ing context-of-query match, visualizing passages within documents, rendering sources and results, reflecting time costs of interaction, managing multiple-search processes, integrating multiple search and browsing techniques, and visualizing large information sets. moreover, zaphiris et al. generalized these tasks into three essential ones: searching, navigation, and browsing. indeed, most infor- mation-visualization projects for digital libraries have emphasized these three tasks. in terms of searching, shneiderman et al. proposed the use of a two-dimensional display with continuous variables to view several thousand search results simul- taneously. this visualization included two strategies: two-dimensional visualizations, and browsers for hier- archical data sets (implemented by using categorical and hierarchical axes). in combination with a grid display, this visualization let users see an overview by color- coded dots or bar charts arranged on a grid and orga- nized by familiar labeled categories. users could probe further by zooming in on desired categories or switching to another hierarchical variable. a language-indepen- dent document-classification system, completed by liu et al., provided a search aid in a digital-library environ- ment and helped users analyze the search query results visually. this system used a vector model to calculate the similarities between documents and a java applet to display the classification to the user. in terms of navigation, there are also a variety of information-visualization applications. the previous example of two-dimensional display developed by shniederman et al. also contained navigation functions. another example is hascoet’s map interface applied to a digital library. this prototype was associated with summary views in the form of navigation trees and neighbor trees that showed documents related to one focus document. the user interface was composed of maps automatically generated based on the characteris- tics of documents retrieved and a default configuration. users could also modify the configuration of maps and edit maps (classical operations such as cut, paste, move, delete, save and load a view, and expand a view). as for browsing, the use of dynamic queries is a tech- nique that has been employed for some time. ahlberg and shneiderman’s ( ) filmfinder is an early example. users can move several sliders to select query param- eters, and the search results change with the movement of the sliders. this tool can help users browse movie records more easily and cognitively. another technique is query previews, proposed by doan et al. query previews allows users to rapidly gain an understanding of the con- tent and scope of a digital collection. users are presented with generalized previews of the entire database using only the most salient attributes. when they select rough ranges, they will immediately learn the availability of the data for their proposed query. all these applications provide good examples and paradigms to some recent projects. this paper’s discussion of visualization techniques will be based on these three essential tasks—searching, navigation, and browsing. ■ techniques and applications this section presents several recent information-visualiza- tion projects applied to digital libraries. all these applica- tions emphasize searching, navigation, and browsing. gang wan gang wan (wangang @gmail.com) is science librarian, texas a&m university, college station. visualizations for digital libraries article title | author visualizations for digital libraries | wan lvis—digital library visualizer indiana university’s (iu’s) lvis (digital library visualiz- er) project aims to aid users’ navigation and compre- hension of large document sets retrieved from digital libraries. borner et al. developed a prototype of lvis based on the data set in the dido image bank, provided by iu’s department of art history. lvis is a good com- bination of information-retrieval algorithms and visual- ized-search interface. in the information retrieval and classification stage, it adopts latent semantic analysis (lsa) to automatically extract semantic relationships between images. the lsa output feeds into a clustering algorithm that groups images into several classes sharing semantically similar descriptors. a modified boltzman algorithm is then used to lay out images in space. this section will focus on the interface metaphors used to dis- play the results of this classification. two interfaces have been implemented for lvis. a d java applet was used on a desktop computer for details, and a d immersible environment for the cave (cave automatic virtual environment). cave is a virtual real- ity ’ x ’ x ’ theater made up of three rear-projection screens for walls and a down-projection screen for the floor. projectors throw full-color workstation fields ( , x stereo) at hz onto the screens, giving between , and , linear pixel resolution to the surrounding composite image. both d and d interfaces give users access to three levels of detail: they provide an overview about docu- ment clusters and their relations; they show how images belonging in the same cluster relate to one another; and they give more detailed information about an image, such as its description or its full-size version. in the cave environment, users can first enter a virtual display theater that stages the digital library as a cyberspace easter island, presenting gateways to specific subject categories established by the previous classification process. borner et al. used d icons here to encode subject categories (in this case, they actually used a sculptural form of heads inspired by images of the data set). after users “enter” into these head icons, they are transited to a new d spatial metaphor that presents images in the current category. these images, or slides from the digital library, are presented in crystalline struc- tures (figure ). in this environment, each crystal represents a set of images with semantically similar image descriptions. again, physical proximity is used here as a metaphor to encode semantic similarities among images. the forma- tions of the crystalline structures depend on the size of the actual search-result data set. navigation in this space is easy. users can also “walk” through this environment and select images of interest to display a larger- and clearer-size version (as two images shown in figure ). if the larger version is not satisfactory, it can be returned to its previous iconic presentation. uc system: a fluid treemap interface for digital libraries the uc system—the acronym “uc” came from its original (but no longer used) internal name “uplib client”—was developed by good et al. at palo alto research center. it was built on the uplib personal digital-library platform, which provides an “extensi- ble, format-agnostic, and secure document-repository layer.” “personal” here means that the user already has the right to use all of the data objects in the library, and already has local possession of those objects. however, this visualization can be employed in more general digi- tal libraries. the uc system uses continuous and quantum treemap layouts to present collections of documents. continuous treemaps are space-filling visualizations of trees that assign an area to tree nodes based on the weighting of the nodes. in continuous treemaps, the aspect ratio of the cells is not constrained, although square cells are often preferred. quantum treemaps extend this idea by making cell dimensions an even multiple of a unit size. the treemap visualizations provide meaningful overviews of document collections and fast, intuitive navigation among multiple documents in a working set. an important aspect of the interface is the fluidity of navigation. this allows the user to focus on the docu- ments rather than on interacting with the tool. the inter- face allows a user to zoom in on an object with a left-click, and to zoom out when the user clicks on the background; figure . lvis: image crystals and panels information technology and libraries | june however, the combination of a “zoomable” user interface and continuous treemaps leads to a problem: conflicts with aspect ratios. to solve this problem, good et al. proposed to zoom and morph the cell to the window size while leaving the rest of the layout in place. thus the visual disturbance of the display is minimized since only a single cell moves. with respect to searching, this system provides sev- eral methods to filter results. first, its interface includes a mechanism to search for specific content within the documents. as letters are added to the search query, the system increasingly highlights matching documents to immediately indicate matching documents. secondly, the user can also choose to update the view to display only those documents that match the current query. figure gives an example of a user-initiated query process. in the scenario described in figure , the user first enters the search terms ( . ), and interactive highlights then appear for groups with matching documents ( . ). the user presses a button to limit the view to only the matching documents ( . ). finally, the user zooms in on a document and begins reading with an integrated reading tool. ( . ) the uc system also offers a mechanism that allows the user to compare multiple documents. after users retrieve a set of documents through a search, they can press a button to “explode” the documents to pages. they can continue zooming in to a portion of a single docu- ment, and then select a document page to read with the integrated reading tool. in short, the uc system uses treemaps as the primary visual metaphor. it also uses various visualization tech- niques that enhance user interactions, such as zooming, interactively highlighting, exploding, etc. activegraph activegraph is an information-visualization tool de- signed by marks et al. ( ) at los alamos national laboratory (lanl). it aims to provide users with a concise, customizable view of documents in a digital library. in this system, a set of digital-library documents is represented as a data set in a d or d scatter plot. the data set can represent any digital-library objects in vari- ous formats including books, journals, papers, images, and web resources. marks et al. used six visual attributes of the scat- ter plot: the x-, y-, and z-axes, color, size, and shape to encode the bibliographic information of documents in a digital library, including title, author, date of publication, and number of citations. the user can select and adjust these attributes from a control panel on the right-hand side of the screen. thus, activegraph allows users to both view and customize the contents of a digital library. the main visual representation of this tool is a scatter plot. scatter plots have been used to represent large sets of data for a long time. they provide an overview of a data set and show the distribution of data points clearly, revealing clusters and statistical information. hence, these scatter plots make it possible for users to perceive meaningful patterns of the data. an example of using activegraph scatter plots to visualize citation data for postdoctoral researchers at lanl is given in figure . this scatter plot intends to provide users with information, such as the number of times their papers published between and were cited. the visualization is based on the metadata in the lanl digital library. in this scatter plot, the postdoctoral researcher’s last name is mapped to the x-axis and the number of cita- tions is mapped to the y-axis. a pixel of a particular color can provide two pieces of information, for example, by encoding a paper and the subject category of the paper. a group of pixels of a particular size, shape, and color can provide four pieces of information by encoding a paper, the subject category of the paper, whether the paper has been cited, and whether the paper has been read by another user of the collaborative library. from this scatter plot, users can easily learn the citation pattern of these papers. unlike some other scatter plot applications such as homefinder and filmfinder, activegraph uses different filters for queries. instead of filter sliders, it uses filter lists, which consist of selection list boxes, one for each data attribute. these filter lists can provide users with functionality that is important in the context of digital libraries. activegraph allows users to manipulate the display of data in another manner by applying a logarith- mic transformation. as some data sets, such as citation data, can frequently have an exponential distribution, the figure . the search interface of the uc system article title | author visualizations for digital libraries | wan logarithmic transformation can spread the clustered dots more evenly across the scatter plot. other data transformations and visualizations may be important in some cases as well, such as parallel coordi- nates for displaying citation statistics for the same group of researchers at different points in time. the scatter plot is not a new visualization technique. this example, however, demonstrates that by encoding document attributes and designing proper filters, it can be used in a digital library environment effectively and efficiently. d vase museum the previous example, lvis, already introduced the applications of d representations in digital libraries. the d vase museum developed by shiaw et al. at tufts university is another good example of d space meta- phor in digital libraries using a variety of visualization techniques. in this d vase museum, the user can navigate seam- lessly from a high level scatter-plot-like plan view to a perspective overview of a subset of the collection, to a view of an individual item, to retrieval of data associated with that item, all within the same virtual room and with- out any mode change or special command. unlike the traditional digital library, which displays thumbnails and descriptions of vases in the main browser interface, this museum is a d virtual environment that presents each vase as a d object and places it in a d space of a room within a museum. figure gives a wide- angle view of this d museum. in this view, one wall represents the timeline (year bce) and the adjacent wall represents the types of wares (e.g., red figures, black figures). the user can “walk” through this virtual room and look at the vases. the wide-angle view pictured in figure will then be tran- sited to an eye-level view so that the user can probe the objects more clearly. when the user continues “walking” toward an object of interest, secondary information about this vase will appear in the virtual scene. if the user looks closer, the text information becomes clearer. as the user moves farther and farther away, the information becomes less and less visible until it eventually disappears from the scene. if the user clicks on the vase html page, a version of the original html page will be loaded, from which a d model of the vase can be loaded. the user can then rotate this d model on the screen using the mouse to see all the aspects of the vase. the d vase museum is maintained in the background all times. the user can also navigate the room in a perspective view by switching the view port upward toward the ceiling (watching from upside down). the user can then switch the views between a high-level scatter plot and a d perspective view. similarly, in this application, the x- and y- axes are mapped to two attributes of the vases: year and ware. with this seamless blend from a high-level data plot to d objects, the user can navigate without los- ing the point of view or context by just moving within the virtual environment. according to shiaw et al., a usability test has been ad- ministered, in which tasks based on archaeology courses were designed and subjects were asked to perform these tasks in the original traditional digital library and this d museum. the results showed subjects who used the d vase museum performed the tasks percent better and did so nearly three times faster. ■ collaborative visual interfaces to digital libraries collaborative visual interfaces is an ongoing project led by borner at indiana university (iu). borner et al. ( ) proposed the development of a shared d document space for a scholarly community—namely faculty, staff, and students at iu’s school of library and information science. the space will provide access to a collection of various online documents including text, images, video, and software demonstrations. a semantic treemap algorithm has been developed to layout documents in a d space. semantic treemaps utilize the original treemap approach to determine the size (dependent on the number of documents) and layout figure . activegraph scatter plot of citation data for papers authored by lanl postdoctoral researchers information technology and libraries | june of document clusters. subsequently, an algorithm (force directed placement) was applied to the documents in each cluster to place documents spatially, based on their semantic similarity, which was encoded by the physical proximity between two dots. an example of the semantic treemaps is shown in figure . . a d space metaphor was then used to display these documents on the desktop interface, as shown in figure . . in this d space, each document is represented by a square panel textured by the corresponding web page’s thumbnail image and augmented by a short description such as the web page title that appears when the user moves the mouse over the panel. as in other d envi- ronments, users can “walk” through this space to probe documents of interest. upon clicking the panel, the cor- responding web page is displayed in the web-browser interface. users can collaboratively examine, discuss, and modify (add and annotate) documents, thereby converting this document space into an ever-evolving repository of the user community’s collective knowledge that members can access, learn from, contribute to, and build upon. certain usability studies have been performed to determine the influence of panel size and panel density on retrieval performance. results showed that subjects were slightly faster and more accurate if web-page panels are larger and denser. ■ aquabrowser aquabrowser is a fuzzy visualization tool that shows the high-level description of a conceptual space, hiding irrelevant information and displaying information ele- ments in context. it is a generic java applet that can be embedded into any web page. medialab, the developer of aquabrowser, claims that users of aquabrowser can browse through a dynam- ic conceptual space that is continually reshaped to reflect their interests. animations make tran- sitions from one state to another appear more fluid, showing users why and how the information is rearranged. medialab uses the term “word cloud” as the visual metaphor of the aquabrowser interface. but in fact, the primary visual representation is a network of linked words that are distributed in the conceptual space. the search term that the user assigns will display at the center of this network. the physical distance between another term node and this term encodes the relevancy between these terms. the larger and closer the word is to the center of the screen, the greater its relevance to the search term. in contrast, the smaller and more peripherally positioned, the less relevant it is. each of the user’s actions will change and rearrange the distribution and importance of the words, putting those of greater interest closer to the user and those of less interest nearer to the edge of the screen. it also uses colors to encode attributes of terms, such as spelling variations, visited words, and translations. figure shows an example of a search display. this tool has been used by a number of libraries to enhance their online catalog search interfaces. it could be a very useful search aid in digital libraries as well. ■ summary and trends the above applications are just a few examples of infor- mation visualization in a digital-library environment. many other metaphors and techniques, such as per- spective wall, cone tree, document lens, and hyperbolic browser, have been used or can potentially be used to facilitate searching, browsing, and navigating through the maze of information in a digital library. the digital library is an interdisciplinary subject involving several research areas such as information retrieval, multimedia information processing, and clas- sification. all these aspects of digital libraries make information visualizations more complicated in this envi- ronment. therefore, the systems described in this paper have integrated various visualization techniques. figure . a wide-angle overview of the d vase museum article title | author visualizations for digital libraries | wan the examples in this paper, along with many others, show that the d space metaphor has attracted much attention from information-science communities. the combination of d space and virtual reality that can be accessed from web browsers these days is becoming a trend of information visualization for digital librar- ies. this technique gives the user maximum freedom to walk through the library collections, searching and browsing documents. the d visual structures, however, have greater implementations compared with those that are d, since they require more processing power and include more parameters. that is partly why many d visualizations developed in the s are still widely used. for example, both activegraph and d vase museum have employed d scatter plots; both uc system and collaborative visual interfaces have used treemaps. furthermore, it is very important to focus on the actual needs of users. research on any visualization for digital libraries should be based on the detailed analy- sis of users, their information needs, and their tasks. usability tests have been done for some of the above applications, but not for all of them. further research and usability tests are required to determine to what extent a visual interface facilitates the user ’s perception of information. references and notes . r. rao et al., “rich interaction in the digital library,” communications of acm , no. ( ): – . . p. zaphiris et al, “exploring the user of information visu- alization for digital libraries,” the new review of information networking , no. ( ): – . . b. shneiderman et al., “digital library search results with categorical and hierarchical axes,” dl- : th acm digital library conference, san antonio (new york: acm pr., ). . y. liu et al, “visualizing document classification: a search aid for the digital library,” journal of the american society for information science , no. ( ), – . . shneiderman et al., “digital library search results with categorical and hierarchical axes.” . m. hascoet, “using maps as a user interface to a digital library,” sigir ’ , melbourne, australia (new york: acm pr., ). . c. ahlberg and b. shneiderman, “visual information seeking using the filmfinder,” acm chi conference, boston (new york: acm pr., ). . k. doan et al., “query previews for networked informa- tion services,” advanced digital libraries conference (washington: ieee, ). . k. borner et al., “lvis—digital library visualizer,” pro- ceedings, ieee international conference on information visual- figure . . interface to the document space figure . . a semantic treemap of web links figure . a search display in aquabrowser information technology and libraries | june ization, july – , , london, england (los alamitos, calif.: ieee computer society, ), – . . c. cruz-neira et al., “surround-screen projection-based virtual reality: the design and implementation of the cave,” computer graphics (proceedings of siggraph ’ ), vol. (new york: acm siggraph, ), – . . k. borner et al., “lvis—digital library visualizer.” . l. e. good et al., “a fluid treemap interface for personal digital libraries,” jcdl’ , june – , denver (new york: acm pr., ). . w. c. janssen and k. popat, “uplib: a universal personal digital library system,” acm symposium on document engineer- ing (new york: acm press, ), . . good et al., “a fluid treemap interface for personal digi- tal libraries.” . b. b. bederson et al., “ordered and quantum treemaps: making effective use of d space to display hierarchies,” acm transactions on computer graphics , no. ( ): – . . l. e. good et al., “zoomable user interface for in-depth reading,” jcdl’ , june – , tucson, ariz. (new york: acm pr., ) . l. marks et al., “activegraph: a digital library visualiza- tion tool,” international journal on digital libraries , no. (mar. ), – . . ibid. . e. r. tufte, the visual display of quantitative information (cheshire, conn.: graphics pr., ). . h. shiaw et al., “the d vase museum: a new approach to context in a digital library,” jcdl’ , june – , tucson, ariz. (new york: acm pr., ). . ibid. . k. borner et al., “collaborative visual interfaces to digital libraries,” jcdl’ , july – , portland, ore. (new york: acm pr., ). . y. feng and k. borner, “using semantic treemaps to cat- egorize and visualize bookmark files,” visualization and data analysis : – january , san jose, usa (proceedings of spie, v. ) (bellingham, wash.: spie—the international society for optical engineering, ), – . . a. veling, “the aquabrowser—visualization of dynamic concept spaces,” journal of agsi , no. ( ): – . . b. eden, “ d visualization techniques: d and d infor- mation visualization resources, applications, and future,” library technical reports , no. ( ). . e. bertini et al., “visualization in digital libraries,” www .dis.uniromal.it/~delos/docs/ivdls_book_chapter.pdf (accessed jan. , ). information technology and libraries | june book review debra shapiro, editor strategic planning and management for library managers by joseph r. matthews. westport, conn.: libraries unlimited, . xiv, p. $ (isbn - - - ). the reality for most librarians is that, sometime in their career, they will be involved in strategic manage- ment and planning. while library school courses occasionally deal with this topic, it is from a theoretical perspective only. most librarians are promoted or coerced into leadership and management roles, often with little or no training or resources at their disposal to assist them with the transition or change of respon- sibilities. strategic planning is one of those duties assigned to library managers and leaders that often get pushed to the lowest-priority list, mainly because there are few guide- lines and handbooks available in this area. since the publication of donald riggs’s strategic planning for library managers (oryx, ), little atten- tion has been given to this vital topic. matthews’s book attempts to provide information on how to explore strat- egies; demystify false impressions about strategies; how strategies play a role in the planning and delivery of library services; broad categories of library strategies that can be used; and identification of new ways to communicate the impact of strategies to patrons. as the author states in the introduction, the focus of librar- ies has moved from collections to encompass the arena of change itself. finding strategies to enable opera- tion in a fluid environment can mean the difference between relevance and irrelevance in today’s competitive information marketplace. the book is divided into three major sections: ( ) what is a strategy, and the importance of having one; ( ) the value of and options for strategic planning; and ( ) the need to moni- tor and update strategies. the first four chapters make up the first sec- tion. chapters and go through the semantics and the need for strategies, as well as the realities and limitations of strategies. chapter provides brief introductions to schools of strategic thought. these include the design school, the planning school, the posi- tioning school, the entrepreneurial school, the cognitive school, the learning school, the power school, the cultural school, the environmen- tal school, and the configuration school. chapter introduces types of strategies: operational excellence, innovative services, customer inti- macy, and the concept of strategic options. section consists of chapters through and provides information on what strategic planning is, what its value is, process options such as plan- ning alternatives and critical success factors, and implementation. section , comprised of chapters and , focuses on the culture of assessment; monitoring and updating strategies; and tools available for managing the library. two appendixes are provided: one containing sample library strate- gic plans, and another with a critique of a library strategic plan. overall, the book is very straight- forward and understandable, with numerous illustrations, process work- flows, and charts. i found the infor- mation very interesting and useful, and the final section on assessment and measurement of strategic plan- ning is essential for libraries to implement and monitor in today’s marketplace. the various explana- tions related to schools of strategic thought were especially helpful. this book should be read by every library manager and director involved in strategic planning and process.—brad eden, associate university librarian for technical services and scholarly communication, university of california, santa barbara ebsco cover lita , , covers and index to advertisers this paper discusses some of the problems associated with search and digital-rights management in the emerging age of interconnectivity. an open-source system called context driven topologies (cdt) is proposed to create one global context of geography, knowledge domains, and internet addresses, using centralized spatial databases, geometry, and maps. the same concept can be described by differ- ent words, the same image can be interpreted a thousand ways by every viewer, but mathematics is a set of rules to ensure that certain relationships or sequences will be pre- cisely regenerated. therefore, unlike most of today’s digital records, cdts are based on mathematics first, images sec- ond, words last. the aim is to permanently link the highest quality events, artifacts, ideas, and information into one record documenting the quickest paths to the most relevant information for specific data, users, and tasks. a model demonstration project using cdt to organize, search, and place information in new contexts while protecting the authors’ intent is also introduced. ■ statement of the problem human history is composed of original events, artifacts, ideas, and information translated into records that are subject to deciphering and interpretation by future gener- ations (figure ). it’s like putting together a puzzle, except that each person assembling bits and pieces of the same information may end up with a different picture. we are at a turning point in the history of humanity’s collective knowledge and expertise. we need more precise ways to structure questions and more interactive ways to interpret the results. today, there is nearly unlimited access to online knowledge collections, information ser- vices, and research or educational networks to preserve and interpret records in more efficient and creative ways. there is no reason digital archiving and dissemination techniques could not also be used to streamline redun- dancies between collections, build cross-references more methodically. content should be presented and tech- niques utilized according to orderly specifications. this will help to document work more responsibly, making shared records more correct, interesting, and complete. the open-source system proposed, context driven topologies (cdt), packs and unpacks ideas and informa- tion in themes similar to museum exhibitions using speci- fications created by each author and network. data layers are formed by registering unique combinations of geogra- phy, knowledge domains, and internet addresses to create multidimensional shapes showing where data originate, where they belong, and how they relate to similar infor- mation over time. the topologies can be manipulated to consolidate and compare multiple sources to identify the most reliable source, block out repetitious or irrelevant background information, and broadcast precise combi- nations of ideas and information to and from particular places. “places,” in this sense, means geographic region and cultural background, knowledge domain and educa- tion level, and all of their corresponding online resources. modern information must be searchable on mul- tiple and simultaneous levels. today’s searches occur for a number of reasons that did not exist when most current collections, repositories, and publications were created. digital records have the potential to reach far broader audiences than original events, artifacts, and ideas. therefore, digitized items and the acts of publish- ing and referencing over networks could theoretically serve a longer-term and more expanded purpose than most individual collections, repositories, or publications are designed to serve. there is no shortage of interesting work to look at. we live in a complex world that is just recently being digi- tized, mapped, analyzed, and broadcast over the internet in fine detail and compelling overall relationships. many deborah l. macpherson (debmacp@gmail.com) is projects director, accuracy&aesthetics (www.accuracyandaesthetics.com) in vienna, virginia. deborah l. macpherson digitizing the non-digital: creating a global context for events, artifacts, ideas, and information digitizing the non-digital | macpherson figure . word word-search-puzzle (courtesy of kevin lightner) information technology and libraries | june of these relationships require mathematics, images, and maps to explain them. we need more than keywords to explore and reference all that has been documented, but we have formed the habit of using keywords and machine-based classification schemes. the entire digital world is in a mire of conflicting priorities, funding oppor- tunities, and intellectual quests toward the future. to advance humanity’s collective curiosity and knowledge, and to coordinate similar efforts across disciplines and cultures, we need one form of record keeping. one global context to show: . where ideas and information begin; . if the original is non-digital (e.g., an artifact or real world event), and if so, the location where the arti- fact resides or the time and place of the event; and . a marking system to keep track of the ways infor- mation has been exchanged, reinterpreted, and reused to create a more comprehensive and simpli- fied guide to humanity’s collective knowledge and expertise. digitizing the non-digital is a concept to address three issues: ■ tools to assemble the bigger pictures needed to docu- ment the best paths to the most relevant information in sets rather than retrieving results item by item; ■ placeholders for information that has not been digi- tized or was never recorded; and ■ distribution to and from specific places according to the ways it is used, the kind of information it is, and the types of people who are able to understand it. there is currently little distinction between all data that have been collected or exist, versus the data and techniques selected to draw conclusions. there are no tools to differentiate between information under rigorous discussion by a discipline or culture versus random bits and pieces. there is a need to develop the equivalent of interpretive exhibits to instruct and inspire the general public. there is currently no way to herd information into crowded areas to be consolidated, compressed, and prioritized by its relationship to similar ideas and infor- mation. citation patterns are able to show connections or structure-related information. however, they currently do not show whether the reference is for or against the other work. there are very few big pictures. there is no way to trace where an idea has led over time. the global context proposed is not like the ancient library of alexandria or large-scale contemporary initiatives. the envisioned process looks beyond the quest to digitize or publish every available event, artifact, and idea. it is not about each item itself. it is being able to make sense of the ways the same information can be viewed in different contexts, and being able to construct a reliable process to search and document the results. having bigger pictures will allow researchers, curators, and others to see what is missing or decide which archival works should be converted into digital form. we do not have the time, resources, or reasons to digitize every item in every collection. the aim is to gradually identify what the most telling examples are in different areas so someone new to an event, artifact, idea, or information can see it in various contexts and automatically be shown the most compelling or instructive sequences first (figure ). a coordinated effort to overlap and see all archives and publications by ranking accuracy and appeal to the public in relationship to all knowledge will make it pos- sible for entirely new lines of inquiry to be established. it will help researchers coordinate work across disciplines. an example of this principle today is the international virtual observatory alliance (ivoa). ivoa is a coor- dinated effort by astronomers worldwide to document our universe more efficiently by systematizing their records; showing where they originate; indicating how they were collected; meeting their rigorous mathematical figure . photomosaic ® thousands of miniature images of the civil war combine to make one large portrait. (courtesy of robert silvers) article title | author digitizing the non-digital | macpherson standards; and deciding themselves how and where their records belong in relationship to each other, and which ones are most important. only astronomers are qualified to do this. the same is true in any area of humanity’s specialized knowledge and expertise. the most difficult aspect of creating a global context is accommodating and expressing each area in its unique way as created from within, while still being able to get the most descriptive examples from all areas to fit together in a sensible and appealing overview. until digital archives and publications can be deeply searched on a global level using simpler tools and prede- termined pathways accessible by anyone, two research- ers in different geographic or academic areas may be investigating the same topic from different points of view and will not know it. there is no way to be led to the best internet resources. today, as so much information surrounds us, it is hard to believe that common lines of inquiry could be discovered by accident. context of the place, time, idea, or education level should be able to drive internet topologies to the most appropriate online resources. constructing a reliable and beautiful digital history of all events—both natural and man-made—artifacts, ideas, and information means contributing to and com- bining a wide range of knowledge, expertise, networks, archives, and tools. mapping digital knowledge to his- torical knowledge means arguing about and perfecting an entirely new set of checks and balances. historical and digital knowledge are different. historical knowledge is fluid, continuous, and held by traditionally separated cultures and disciplines. digital knowledge goes every- where that can be marked and traced by the times and places it was created, captured, and distributed. trying to visualize what is happening and relating it to working practices and the types of information that came before it is not like tracing the history of the human race back to adam and eve or the universe back to the big bang, where substantial guesswork beyond our memory or experience is involved. the entire conversion into the networked age is happening before our eyes in less than one generation without the benefit of reflection, care- ful review, and storytelling. we’re collecting everything indiscriminately over and over again while all datasets are rapidly expanding. we need to step back, slow down, and acknowledge that many current digitization and publication methods do not consistently generate reflec- tive or reviewed results that are able to tell a story. we do not currently have one shared map, context, mathematical record, language, or set of symbols to interpret from different points of view for a variety of purposes over time. we do not currently mark the origi- nal versus subsequent interpretations of the same infor- mation as an integral component of most digital records. there is no financial support for one single shared stor- age space to preserve only the highest-resolution, most agreed-upon versions because we may never be able to agree on what they are. therefore, there is also not one system that can be fine-tuned to discover research and results that may be accidentally overlapping. instead, unusual approaches get watered down by constrained words designed to fit metadata requirements devel- oped by archivists and engineers rather than the origi- nal authors. links get broken, web sites are no longer maintained, trends change. there are currently very few feasible ways to pick up on a line of inquiry previously initiated by others without sorting through and regener- ating the same information again. a simplified version of the work needs to be preserved on the network, able to be referenced by others even if they are far away, live in a different time, or are more or less advanced in their ways of thinking. if digital information is reliable, someone in a remote place or in the future should not need to collect the same information again or unintentionally retrieve out-of-date or duplicate results. searches in the public domain should not be boring. they should be as easy to click through as tv chan- nels, with more directions to go and better content. all searchers should not have to start at the top like everyone else on the first page of google, citeseer, or arxiv with a blank white space and a box to enter key words. investigators should be able to outline the facts they know, dial in measurements, specify relationships, and generally be able to use their own knowledge and expertise to isolate and extract entire ideas over broad spectrums or select only relevant portions of archives and publications to reintegrate into larger bodies of work for further discussion. digital objects are able to depict more than the unaided eye can see. an example is the evaluation of the center of mass of michelangelo’s david performed for david’s restoration by the visual computing lab based on a d model of the statue built by stanford university (figure ). the digital david does not have mass. the original david is a beautiful object sculpted of a known and pre- dictable material. the model makes it possible to test res- toration techniques without permanent damage in ways no one would dare attempt on the irreplaceable original without first knowing more. the documentation process is an enhanced original that should be permanently bound to the digital history of the original sculpture. the evaluation method could be applied to other objects, but this model belongs with this object and this type of research. a global context built upon a solid, mathemati- cally linked foundation would mean this conscientious work would not be lost or need to be repeated. digital records are not being used nearly to their full potential. so many influences on humanity’s intellectual evolution could be examined as history takes shape over information technology and libraries | june time. concurrent and conflicting interpretations can take on more meaning than the original by itself. for example, how could the internet and legal citations be used to map the subsequent interpretations of the u.s. constitution from the time, place, and reasons where it was written to every supreme court case and related citation since the original context? what would this map look like (figure )? the impact that these four pages of ink on paper have had to the united states and the entire world can- not currently be examined in one volume to see where the most contentious and useful passages are. similar dynamics in wikipedia are shown in history flow by martin wattenberg at ibm research. what if techniques developed in one field could be applied to content from another area? for example, what if computer models created to track storms and hurricanes could be used to arrange and watch the evolution and real world impact of all the documents and actions associated with a war? being able to see how originals evolve in their inter- pretation and impact on society over time is practical because not all records are worth keeping. even worse, mundane or meaningless events, artifacts, ideas, or infor- mation may seem more important than they actually were if they are not translated into digital form or distrib- uted in the right way. the task today is to make the most advanced ways of thinking and working more approachable and appealing to someone new, which is everyone outside a particular discipline or culture, while traversing a map of humani- ty’s collective knowledge and expertise. because shared memories of this magnitude would be so far-reaching and complex, the record itself needs to be able to show every user how to use it. every unique purpose for look- ing around, publishing, or referencing work, and adding to or taking away from a collaborative global context should be geared toward improvement and simplifica- tion. while millions and millions of people are accessing enormous numbers of files and collections, some paths are better than others. in order to sort and choose the best parts of vast collections, documenting everyone going in and out of various semantic places can ultimately iden- tify the best paths to information everyone understands. what if someone who does not care at all about paintings makes an inquiry—which ten should they be shown to get them interested? there is also the issue of gearing the internet to provide more efficient pathways to widely accessed preapproved and curated information. every mouse click could accumulate to document the most reliable pathways in and out of shared information spaces to generate an assortment of scenarios for looking at the same information in different ways (figure ). we think there is far too much information to con- solidate into one big picture, that our ideas and methods are too incompatible to coexist comfortably in one space, but perhaps this is not really the case. perhaps we can understand what is happening more clearly by working backwards. ■ proposed solution and design for a running prototype even though many networks are in place and count- less computers have been manufactured, technology advances rapidly. there are very few reasons to repair obsolete equipment or maintain outdated web resources. therefore, why not go back to the drawing board on all of it? we may have completely new computers and net- works within ten years, anyway. a record-keeping and referencing system this ambi- tious needs to incorporate every type of record, classifica- tion scheme, symbol, style, and quirk. when visiting a new place outside your comfort zone, it needs to be obvious what the best local techniques are to filter and understand the results. people new to an area need to have the option of using tools they can invent or already know. figure . david’s center of mass (courtesy of the visual computing lab and stanford university) article title | author digitizing the non-digital | macpherson the visualization of cdt’s model demonstra- tion project will bring together research scien- tists, artists, integrators, and institutions to devel- op a running prototype. the purpose is to establish and record a series of planned and spontane- ous situations in different parts of the world across a range of disciplines and existing networks so that these situations can be mapped. the project will be a group of people thinking together to confront the road- blocks in assembling incompatible ideas and information into one context. the group will collaborate in larger and smaller groups in roughly three-month intervals as par- ticipants continue with their existing work. the develop- ment of this system has to be dynamic, changing piece by piece both from the bottom up and the top down while everyone’s regular work continues. therefore, the system will be geared toward sample sets of active work products, rather than the record-keeping system by itself. the current objective is to establish a network of ten art museums, ten scientific research institutes, and ten new media/new technology efforts in ten cities that speak different natural languages (for example: english, german, french, italian, hindi, mandarin, ga [belong- ing to the cluster of kwa languages in ghana], uzbek, spanish, and arabic). the overall intent is to use math- ematics, art, and individual ways of knowing to develop a series of professional sketches to serve as shortcuts between languages and key words in the search process. the first step is to map the background of each of the project participants’ previous work by time, location, and discipline. the database will include scientific visu- alizations, art objects, performances, algorithms, math- ematical formulae, musical recordings, and many other forms of creative and scholarly expression. the next steps will be to hold a series of interactive workshops. at the first workshop, the research scientists will explain the mathematics and images they use in their work. two sets of artists will isolate the aesthetics to render their own map through the scientists’ ideas. two traveling exhibits will be created, one to be experienced in per- son, the other to be presented through a new media and online exhibit. both will be tracked physically and con- ceptually using cdt. the results will be generated and interpreted using gis, matlab, photoshop, and flow visualization software. for more information, please contact the author. a survey of individual and institutional requirements will be undertaken to define practical ways to move and organize ideas and information into a unified sample map of previously unrelated content and techniques. for example, at one institute, perhaps only two participants and four local professors will understand what that part of the map is showing. another part may only have meaning to one artist. a unified map for everyone, with built-in copyright protection for the participating artists, scientists, and institutions, will be presented to nonspe- cialist general publics around the world for feedback and further change within specified limits. the participating publics will be people interested in contemporary art, cutting-edge scientific research, new media, and events where all three communities can interact. each part of the prototype will be able to be examined in groups to compare and contrast different elements against different backgrounds. some arrangements will be assisted by the computer and network. the project will map everything with which each event, idea, and artifact has ever been associated in scale, proportion, and relative placement in the record overall. for example, if the records in question are paintings, any group could figure . thick and thin (courtesy of the artist john simon) figure . the constitution of the united states (courtesy of the u.s. national archives and records administration [nara]) information technology and libraries | june be gathered together into the same reference window without copying the images. the assembly window has a built-in scale for the items it is showing, so they will be displayed in the correct proportion to each other. the system binds images of physical objects with their dimen- sions and the times and places they were created while this information is known—so a user does not ever have to guess later when looking back at any part of the record. any group of paintings can be automatically arranged chronologically, by size, culture, or any number of com- parisons and curatorial issues. a sample sequence is: . a zoomed-in map showing a group of paintings in an exhibit. each painting links to its history. . within the map of all paintings shown in an intri- cate collage. . inside the map of all human endeavor shown as an appealing landscape. higher levels can then be used to reorganize a theme, for example, “only germany to ,” and drilling back down to generate other exhibitions. this would lead to other paintings and other curators’ conclusions, which would provide a more complete representation of each painting, exhibition, museum, curator, culture, and era. when the records in question are scientific visualizations, problems of presenting unrelated files together are more complex. the records may not share a common scale or system of reference. it may only be possible to place mathematical constructs in contexts based on where they originate geographically and by knowledge domain. an important part of the work will be determining the best contexts by which to introduce ideas or information to untrained viewers and devising methods to start deeper in the records using mathematical, cultural, or other prior knowledge and preferences. the same concept can be described by different words, the same image can be interpreted a thousand ways by every viewer, but mathematics is a set of rules to ensure that certain relationships or sequences will be precisely regenerated. therefore, unlike most of today’s digital records, cdts are based on mathematics first, images second, words last. ideas and information will be encoded to persist over specified periods of time. better examples will find higher placement by connecting to more background information and showing stronger relationships to larger numbers of open questions. cycles will be implemented to return to the same idea later and remove information that is never referenced or has not changed the course of the record’s flow. out-of-date, irrelevant, or rarely used information has to either be compressed or be thrown away, a new type of identity and a process to assemble and eliminate information will be created in thirty prototype forms showing the intertwined history of the events, artifacts, ideas, and information generated by the project and all it branches out to when connecting back to the publications, exhibits, ideas, artifacts, and other infor- mation generated by the participating individuals and institutions. the cdt model will relate and join tables to display all the different forms together in one map. each piece of information and the patterned space around it will be documented a special way to generate drawings leading back to originals reliably structured to transfer to other computers and networks. they will transfer with- out ambiguity because the transactions and paths to the internet addresses are based on mathematical relation- ships that can be checked. each contributor has the first opportunity to place his or her ideas in context and define the limits of how their originals can be referenced, changed, and presented. at the end of the project, the set will be closed so that it can be cleaned of information that was only temporary, place- holders can be examined, and the entire model can be manipulated as one whole. for more information, please see www.contextdriventopologies.org the more specified a single piece or set of informa- tion, the easier it will be to define its history and place it in context. each unique placement and priority assigned by each individual or institution may not agree with the priorities and placements envisioned by others, but sooner or later, there will begin to be correspondence and everyone will be looking very generally at the same emerging map. ■ conclusions there will be innumerable contexts to create, discover, and remark upon in the future by creating a shared pace of curiosity and knowledge acquisition. a global context could be used to extrapolate new knowledge from trends that occur over longer periods of time in more places than we currently share or document. as the envisioned system is fine-tuned, it will become an ideal place to test an idea that is only partially complete to see where the idea fits or to determine if it has already been done. the results could be immediately applied to improve educa- tion. in today’s frantic information overload, we should not forget that digital information—and even cold, hard, raw data—is more than ones and zeros. they represent peoples’ work, their fingerprints; people are attached to their data. one wishes networks of computers could understand one’s ideas and work, but we only show them the boring parts. the proposed system will capture beauty so com- puters can help to find where it is hidden inside all the repositories, publications, and collections through which no person has the time to sort. the system will allow article title | author digitizing the non-digital | macpherson users to specify how they think their information relates to the rest of the world so their intended context can be traced in the future. one hopes that using networks and computers to compare ideas and works on larger levels will restore craftsmanship and attention spans to make users want to spend more time with better information. a shared visual language driven by mathematical relationships that can be checked will allow future his- torians to see where records simply will not harmonize. users will be able to analyze why different ways of look- ing can shape and divide knowledge and history as it changes. visiting online archives and publications will change. developing processes to pre-organize searches and results for public viewing can change now by creat- ing a system for curators and others to develop sets of information, rather than publishing individual items on their web sites. library facilities can change, and research rooms can become multimedia centers. networks can broadcast content and techniques in one package. there is not one clearly defined reason why being able to see these kinds of overviews or make these types of comparisons can be useful. the internet is a worldwide invention being constructed for a variety of purposes. a perfectly legitimate reason to capture the history of trans- actions across it in a simple form is just to see what might happen with the objective of increasing our understand- ing and respect for each other. the most important reason for establishing a global context is to allow users to trans- fer and update complex histories, thoughts, images, stud- ies, visualizations, drawings, flow diagrams, sequences, transformations, cultural objects, stories, expressions, and purely mathematical or dynamic relationships without depending on constrained keywords or illegible codes that do not describe this information as well as the infor- mation can describe itself. all cultures and disciplines would be able to construct their parts of the record pre- cisely the way they prefer. we would finally be able to use computers to show why and how we think information is related—a huge leap forward in the world of digital record keeping. references and notes . citeseer, , http://citeseer.ist.psu.edu (accessed apr. , ); internet , , www.internet .edu (accessed apr. , ); jane’s information group, , www.janes.com (accessed apr. , ); machine learning network online information service (mlnetois), , www.mlnet.org (accessed apr. , ); national technical information service, , www.ntis .gov (accessed apr. , ); smithsonian institution librar- ies, galaxy of knowledge, , www.sil.si.edu/digitalcollec- tions (accessed apr. , ); thompson scientific, isi web of knowledge, , www.thomsonisi.com (accessed apr. , ); visual collections, david rumsey collections, , www .davidrumsey.com/collections (accessed apr. , ); world health organization, statistical information system, , www .who.int/whosis/menu.cfm (accessed apr. , ). . g. ammons et al., “debugging temporal specifications with concept analysis,” in proceedings of the acm sigplan conference on programming language design and implementation (new york: association for computing machinery, june ). . w. huyer and a. neumaier, “global optimization by multilevel coordinate search,” in global optimization ( ): – . a. bagga and b. baldwin, (workshop paper), in coling- acl ‘ : th annual meeting of the association for computational linguisitics and th international conference on computational linguisitics, aug. – , , montréal, quebec, canada: proceed- ings of the conference (new brunswick. n.j.: acl; san francisco: morgan kaufmann, ); s. deerwester et al., “indexing by latent semantic analysis,” journal of the american society for information science , no. ( ): – ; a. mccallum and b. wellner, “toward conditional models of identity uncertainty with application to proper noun coreference,” in proceedings of the ijcai workshop on information integration on the web (moun- tain view, calif: research institute for advanced computer science, ), – ; t. nisonger, “citation autobiography: an investigation of isi database coverage in determining author citedness,” college and research libraries , no. (mar. ): – ; k. van deemter and r. kibble, “on coreferring: corefer- ence in muc and related annotation schemes,” computational linguistics , no. (dec. ); k. boyack, “mapping all of science and technology at the paper level,” presented at the session mapping humanity’s knowledge and expertise in the digital domain as part of the st annual meeting of the association of american geographers (denver: association of american geogra- phers, ): ; metacarta, , www.metacarta.com. . j. burke, “knowledgeweb project, .” www.k-web .org (accessed apr. , ); visual browsing in web and non-web databases, iowa state university, www.public.iastate .edu/~cyberstacks/bigpic.htm (accessed apr. , ). . international virtual observatory alliance, , www .ivoa.net (accessed apr. , ). . s. bradshaw, “charting excursions through the literature to manage knowledge in the biological sciences,” presented at the session mapping humanity’s knowledge and expertise in the dig- ital domain, as part of the st annual meeting of the association of american geographers (denver: association of american geog- raphers, ): , project paper available from http://dollar .biz.uiowa.edu/~sbradsha/beedance/publications.html (accessed apr. , ). . m. callieri et al., “visualization and d data processing in the david restoration,” ieee computer graphics and applica- tions , no. (mar./apr., ): – . . m. wattenberg, “history flow,” , http://research web.watson.ibm.com/history (accessed apr. , ). . k. börner, “semantic association networks: using semantic web technology to improve scholarly knowledge and expertise management,” in visualizing the semantic web, nd ed. vladmire geroimenko and chaomei chen, eds., (london: springer verlag, ) – . . g. sidler, a. scott, and h. wolf, “collaborative brows- ing in the world wide web,” in proceedings of th joint european networking conference, edinburgh, scotland (new york: elsevier, information technology and libraries | june ); j. thomas, “meaning and metadata: managing informa- tion in a visual resource reference collection,” in proceedings of association for computers and the humanities and the association for literary and linguistic computing meeting (charlottesville, va.: university of virginia, ); h. yu and a. vahdat, “design and evaluation of a conit-based continuous consistency model for replicated services,” in acm transactions on computer systems , no. (aug. ): – . . visualization of context driven topologies/cdt model demonstration project, , www.contextdriventopologies.org (accessed apr. , ). image acknowledgments: -word word-search puzzle www.synthfool.com/puzzle.gif permission: kevin lightner, synthesizer enthusiast. wrightwood, california abraham lincoln www.photomosaic.com/samples/large/abrahamlincoln.jpg permission: from the artist robert silver. david’s center of mass http://vcg.isti.cnr.it/projects/davidrestoration/restaurodavid.htm http://graphics.stanford.edu/projects/mich/book/book.html permission: roberto scopigno, visual computing lab, isti-cnr, via g. moruzzi, , pisa italy and marc levoy, stanford computer graphic lab, gates computer science bldg. stanford, ca u.s. constitution www.archives.gov/ repository: national archives building, washington, d.c. permission: nara government records are in the public domain. thick and thin www.numeral.com/drawings/plotter/thickandthin.html " × " ink on paper. permission: from the artist john simon, new york city. specializing in algorithms and conceptual art. antelman information technology and libraries | september article title: subtitle in same font author name and second author author id box for column layout library catalogs have represented stagnant technology for close to twenty years. moving toward a next-gen- eration catalog, north carolina state university (ncsu) libraries purchased endeca’s information access platform to give its users relevance-ranked keyword search results and to leverage the rich metadata trapped in the marc record to enhance collection browsing. this paper discusses the new functionality that has been enabled, the implemen- tation process and system architecture, assessment of the new catalog’s performance, and future directions. editor’s note: this article was submitted in honor of the fortieth anniversaries of lita and ital. t he promise of online catalogs has never been realized. for more than a decade, the profession either turned a blind eye to problems with the catalog or accepted that it is powerless to fix them. online catalogs were, once upon a time, “the most widely-available retrieval system and the first that many people encounter.” needless to say, that is no longer the case. libraries cannot force users into those “closed,” “rigid,” and “intricate” online catalogs. as a result, the catalog has become for many students a call-number lookup system, with resource discovery happening elsewhere. yet, while the catalog is only one of many discovery tools, covering a proportionately narrower spectrum of information resources than a decade ago, it is still a core library service and the only tool for accessing and using library book collections. in recognition of the severity of the catalog problem, particularly in the area of keyword searching, and seeing that integrated library system (ils) vendors were not addressing it, the north carolina state university (ncsu) libraries elected to replace its keyword search engine with software developed for major commercial web sites. the software, endeca’s information access platform (iap), offers state-of-the-art retrieval technologies. ฀ early online catalogs larson and large and beheshti summarize an extensive body of literature on online public access catalogs (opacs) and related information-retrieval topics through . the literature has tapered off since then; however, as promising innovations failed to be realized in commercial systems, mainstream opac technology stabilized, and the library community’s collective attention was turned to the web. first generation online catalogs ( s and s) provided the same access points as the card catalog, dropping the user into a pre-coordinate index. the first online catalogs, byproducts of automating circulation functions, were “intended to bring a generation of library users familiar with card catalogs into the online world.” the expectation was that most users were interested in known-item searching. with the second generation of online catalogs came keyword or post-coordinate (boolean) searching. while systems based on boolean algebra represented an advance over those that preceded them, boolean is still a retrieval technique designed for trained and experi- enced searchers. (twenty years ago, salton wrote, “[t]he conventional boolean retrieval methodology is not well adapted to the information retrieval task.” ) boolean systems were, however, simple to implement and eco- nomical in their storage and processing requirements, important at that time. soon after the euphoria of combining free-text terms across records wore off, the library community recognized that the major problem with first- and second-generation catalogs was the difficulty of searching by subject. ฀ the “next-generation” catalog by the early s, thinking turned to next-generation catalog features. out of this surge of interest in improv- ing online catalogs emerged a number of experimental catalogs that incorporated advanced search and match- ing techniques developed by researchers in information retrieval. they typically did not rely on exact match (boolean) but used partial-match techniques (probabilistic and vector-based). since probabilistic and vector-based models were first worked out on document collections, not collections of marc records, adaptations were made to the models. these prototype systems included okapi, which implemented search trees, and cheshire ii, which refined probabilistic retrieval algorithms for online cata- logs. it is particularly sobering to revisit one system that was developed between and . the cite catalog, developed at the national library of medicine, incorpo- rated many of the features of the endeca-powered catalog, including suggesting (mesh) subject headings, correcting spelling errors, stemming, as well as even more advanced features, such as term weighting, keyword suggestion, and “find similar.” toward a twenty-first century library catalog kristin antelman, emily lynema, and andrew k. pace kristin antelman (kristen_antelman@ncsu.edu), emily lynema (emily_lynema@ncsu.edu), and andrew k. pace (andrew_pace@ncsu.edu) are respectively associate director for the digital library, systems librarian for digital projects, and head, information technology, at the north carolina state university libraries, raleigh. toward a twenty-first-century library catalog | antelman, lynema, and pace ฀ where are we now? as belkin and croft noted in , “there is a disquiet- ing disparity between the results of research on ir tech- niques . . . and the status of operational ir systems.” two decades later, libraries are no better off: all major ils vendors are still marketing catalogs that represent second- generation functionality. despite between-record linking made possible by migrating catalogs to web interfaces, the underlying indexes and exact-match boolean search remain unchanged. it can no longer be said that more sophisticated approaches to searching are too expensive computationally; they may, however, to be too expensive to introduce into legacy systems from a business perspective. ฀ the endeca-powered catalog coupled with the relative paucity of current literature on next-generation online catalogs is a scarcity of library industry interfaces from which to draw inspiration, rlg’s red light green and oclc’s fictionfinder being notable exceptions. in june , library automation vendor tlc announced a partnership with endeca technologies for joint sales, marketing, technology, and product develop- ment of the company’s iap software. this search software underlies the web sites of companies such as wal-mart, barnes and noble, ibm, and home depot. ncsu libraries acquired endeca’s iap software in may , started implementation in august, and deployed the new catalog in january . several organizational and cultural factors contrib- uted to making this project possible. of significance was an ongoing administrative commitment to fund digital- library innovation, including projects that involve some risk. library staff share this feeling that calculated risks are opportunities to improve the library as well as to open up new challenges in their own jobs. critically, they also believe that not all issues, particularly “edge cases,” (i.e., rarely occurring scenarios) must be resolved before releasing a new service. finally, it was important that the managers who controlled access to programming and other resources were also the project leaders and drivers of the collective urgency to solve the underlying problem. all these factors also contributed to making possible a five-month implementation timeline. functionality the principle functionality gained by implementing an advanced search-and-navigation technology such as the endeca iap falls in three main areas: relevance-ranked results, new browse capabilities, and improved subject access. most ilss, including ncsu’s former catalog, presented keyword results to users in one order: last-in, first-out (i.e., system sort), while browsing within key- word result sets was limited to the links within individual records. ฀ searching and relevance ranking of results inhabiting the catalog search landscape now, somewhere between a second- and third-generation catalog, is endeca’s mdex engine, which is capable of both boolean and limited partial-match retrieval. queries submitted to endeca can use one of several matching techniques (e.g., matchall, matchany, matchboolean, matchallpartial). the current ncsu implementation primarily uses the “match- all” technique for keyword searching, an implied and technique that requires that all search terms (or their spell- corrected, truncated form) entered by the user occur in the result. the user is not required to enter boolean operators for this type of search; in fact, these terms are discarded as stopwords. the “matchboolean” technique continues to support true boolean queries with standard operators; access to this functionality is provided through advanced search options. although classic information retrieval research tends to associate relevance ranking with probabilistic or vec- tor-based retrieval techniques, endeca includes a suite of relevance ranking options that can be applied to boolean- type searches (i.e., implied and/or). these individual modules are combined and prioritized according to cus- tomer specifications to form an overall relevance ranking strategy, or algorithm. each search index created in the endeca software can be assigned a different relevance ranking strategy. this capability becomes significant when considering the differences in the data being indexed for isbn/issn as compared to a general keyword search. since the keyword anywhere index contains the majority of the fields in a marc record and is the default search operator, its rel- evance ranking strategy received the most attention. this strategy currently consists of seven modules. the first five modules rank results in a dynamic fashion, while the final two modules provide static ordering based on publication date and total circulation. the ncsu libraries, algorithm prioritizes results with the query terms exactly as entered (no spell-correction, truncation, or thesaurus matching) as most relevant. for multiterm searches, results containing the exact phrase are considered more relevant than those that do not. in addition, ncsu has created a field priority ranking, which information technology and libraries | september provides the capability to define matches that occur in the title as more relevant than matches that occur in the notes fields. the relevance algorithm also considers factors such as the number of times the query appears in each result and the term frequency/inverse document frequency (tf/idf) of query terms. the unprecedented nature of using this particular set of tools to define relevance algorithms in library catalogs meant that the initial configuration required a best guess approach. the ability to quickly change the settings and re-index provided the opportunity both to learn by doing and test assumptions. much work remains, however, including systematic testing of the “matchallpartial” retrieval technique. while not a true probabilistic or vector- based matching approach, the “matchallpartial” retrieval technique will broaden a search by dropping individual query terms if no results are returned. however, this type of retrieval technique creates the challenge of developing an intuitive interface that helps users understand partial matching (although many users must be aware that this is how google works). spell correction, “did you mean . . . ,” and sort several other features are included in the basic endeca iap application. these include auto-correction of mis- spelled words, which uses an index-based approach based on frequency of terms in the local database rather than a dictionary. due to the presence of unique terminology in the database (particularly author names), the relevance ranking has been configured to display any matches on the user’s original term before spell-corrected matches. a “did you mean…” feature also checks queries against terms indexed within the local database to determine if another possible term has more hits than the original term in order to provide the user the option to resubmit the search with a different spelling. various sort options are supported, including date, title, author, and “most popular.” ฀ browse whatever the shortcomings of the card catalog, a library user could approach it with no query in mind; any drawer could be browsed. with the advent of online catalogs, this is no longer possible: an initial search is required to enter the system. marchionini characterizes “browsing strategies” as “informal and opportunistic.” a good catalog browse should simulate the experience of browsing the stacks, even potentially improving upon it since the virtual browser can jump around. many patrons cite the seren- dipity of browsing the stacks and “recognizing” relevant resources as a key part of their discovery process. with more books moving to online formats and off-site storage (and therefore, unable to be browsed), enhancing virtual browsing in the catalog becomes increasingly important. as borgman points out, “few systems allow search- ers . . . to pursue non-linear links in the database.” key browsing features provided by the endeca software are faceted navigation and the ability to browse the entire collection without entering a search term. although most modern search engines support both fast response times and relevance ranking, the opportunity to apply endeca’s guided navigation feature to the highly structured marc record data was particularly intriguing. guided, or faceted, navigation exposes the relationships between records in the result set. for example, a broad topi- cal search might return thousands of results. classification codes, subject headings, and item-level details can be used to define logical clusters for browsing—post-coordinate refinement—within the result set. since these refinements are based on the actual metadata of the records in the result set, users can never refine to less than one record, (i.e., there are no “dead ends”).these clusters, or facets, are known as dimensions. users are able to select and remove values from all available dimensions in any order to assist them as they browse through the result set. endeca’s dimensions, while able to be browsed, are not available only as post-coordinate search refinements, however. using the endeca application, library catalogs can once again give users the ability to browse the entire set of records without first entering a search term. any of the dimensions can be used to browse the collection in this fashion, and the ability to assign item-level infor- mation (e.g., format, availability, new book), as well as bibliographic-record elements, to the dimensions further enhances the browsing functionality. ฀ improving subject access given the unsuitability of library of congress subject headings (lcsh) as an entry vocabulary, improving topical (subject) access in catalogs centers around keyword searching. while keyword searches query the subject headings as they do the rest of the record, most systems do not take advantage of the fact that subject headings are controlled and structured access points or use the subject information embedded in the classification number. the endeca-powered catalog, in addition to address- ing classic keyword-search problems by introducing relevance ranking, implied phrase, spell correction, and stemming, also leverages the “ignored” controlled vocabulary present in the bibliographic records—subject headings and classification numbers—to aid in improv- ing topical searching. this is a system design concept that has been discussed in the literature on improving subject toward a twenty-first-century library catalog | antelman, lynema, and pace access but has not until now been manifest in a major catalog implementation. as chan noted, “subject headings and classification systems have more or less operated in isolation from each other.” the endeca-powered catalog interface is an experiment in presenting users with these two different, but complementary, approaches to categorizing library materials by subject. classification several catalog experiments created retrieval clusters based on dewey- and ddc-classification schemes and captions in order to improve subject access by expanding the entry vocabulary and as a way to improve precision and recall. using the lc classification is more challeng- ing, however, as it is not hierarchical. still, the potential of its use has been noted by bates and coyle; and larson experimented with creating clusters (“classification clus- ters”) based on subject headings associated with a given lc class. in larson’s system, the interface suggested possible subject headings of interest, an approach similar to that of displaying the subject facets alongside the result set in the endeca catalog. there is some evidence from early usability studies that exposing the classification, much as it was physically exposed in the card catalog, is useful and desired by catalog users. markey summarizes findings of a council on library resources study in which many institutions con- ducted usability testing. positive aspects of card-catalog use that people wanted to see in the opac included, a “visual overview of what is available in the library,” and “serendipity.” but there is a difference between using the classification scheme to identify subject headings and displaying the classification itself in the user interface. the latter can be problematic from a usability perspective, as larson pointed out, because the classification scheme and terminology are not transparent. imagine the would-be browser of a library’s computer-science collection hav- ing to know to select first q science, then qa –qa mathematics, and then qa –qa instruments and machines before possibly recognizing that qa –qa . calculating machines included computer science? despite these potential problems, because the endeca software supported display of the lc classification as a dimension, ncsu decided to experiment with its utility by making it available on the results screen. entry vocabularies entry vocabularies or mappings apply to all types of retrieval models. they address the general problem of reconciling a user’s query vocabulary with the index vocabulary represented in the catalog or documents. studies show that users’ query vocabulary is large (people rarely pick the same term to describe the same concept) and inflexible (people are unable to repair searches with synonyms.) because of this, bates refers to the objective of the entry vocabulary as the “side-of-a- barn principle.” several approaches have been taken to develop this functionality. building on larson’s “classification cluster- ing” methodology, buckland created an entry vocabulary module by associating dictionaries created by analyz- ing database records. the result was natural language indexes to existing thesauri and classification systems. while the endeca-powered catalog does not yet incorporate an entry vocabulary, its exposure of the index vocabulary to the user in subject dimensions could be said to be a limited side-of-a-barn approach. the limitation is that only controlled vocabulary from the retrieved records is exposed as dimensions on the results screen; relevant records not retrieved because of a lack of match between query vocabulary and terms in the record will not have their facets displayed. were an entry vocabulary for lcsh available, endeca’s synonym-table feature could be used to map between query terms and lcsh. ฀ implementation the library’s information technology advisory committee appointed a seven-member representative team to oversee the implementation. preparatory steps included sending key development staff to training and a two-day meeting with endeca project managers to establish functional and technical requirements. architecture knowing that the endeca application would not com- pletely replace ncsu’s integrated library system, deter- mining how best to integrate the two products was part of the implementation process. the endeca iap coexists with the sirsidynix unicorn ils and the sirsidynix (web ) online catalog, indexing marc records that are exported from unicorn’s underlying oracle database. figure depicts the integration of the endeca software with exist- ing systems. although the endeca software is capable of communicat- ing directly with the database that supports the unicorn ils, ncsu chose the easier path of exporting marc records into text files for ingest by endeca. the marc j api is used to reformat the exported marc records (which include item- level information in fields) into flat text files with utf- encoding that are parsed by endeca’s data foundry process. nightly shell scripts export updated and new records from ils, merge those with the base endeca files, and start the re-indexing process. the indexing of seventy-three marc information technology and libraries | september record fields and ten dimensions results in an index size of approximately . gb. the entire index resides in system memory. the endeca data foundry can easily parse and re- index the approximately . million titles in ncsu’s holdings nightly (in stark contrast to the more than days of down- time required to re-index keywords in unicorn). the relative speed of this process and the fact that it does not interfere with the front-end application prompted the decision not to implement “partial indexing” at the outset. though there was little doubt among staff as to the increased capabilities of keyword searching through endeca, the implementation team decided that authority searching (author, title, subject, call number) would be preserved in the new catalog interface. this allowed ncsu to retain the value of authority headings, in addition to providing a familiar interface and approach to known-item searching. since the detailed record in web included the capability to save records, place requests, and send system- suggested searches (“more like this”), the implementation team also decided to link from titles in the endeca-pow- ered results page to the web detailed record. only slight modifications were required to stylize this display in a manner consistent with the new interface. the front-end interface for keyword searching in endeca is a java-based web application built in-house. this application is responsible for sending queries to the endeca mdex engine—the back-end http service that processes user queries—and displaying the results that are returned. user-interface design because it is created by the customer, ncsu libraries has complete control over the look, feel, and layout of the endeca search-results page. indexes, properties, and dimensions the implementation team began the process of making indexing decisions by looking at the fields indexed in the unicorn keyword-index file. this list included marc fields and subfields, including more than thirty fields that are never displayed to the public. this kitchen-sink approach was replaced with a more carefully selected list less than half that number. the implementation team defined eleven dimensions for use with endeca’s faceted navigation feature. once users enter a search query, they can explore the result set by selecting values from these dimensions: availability; lc classification; subject: topic; subject: genre; format; library; subject: region; subject: era; language; and author (see figure ). the eleventh dimension is not dis- played on the results page, but is used to enable patrons to browse new titles. each dimension value also lists the number of results associated with it; most dimensions are listed in frequency order. search interface once the implementation team made some preliminary decisions regarding dimensions and search indexes, wire- frames were created to assist in the iterative design process for the front-end application. while the positioning of the dimensions on the results page and the display of holdings information was well debated, the design of the catalog search page was an even hotter topic. integration of both endeca keyword searching and web authority searching required an interface that could help users differentiate between the two tools. a survey of the keyword-versus-authority search- ing distinction in a variety of library catalogs led to the development of four mock-ups. the implementation team chose a search tab that includes separate search boxes for keyword and authority searching, as well as search figure . ncsu endeca architecture figure . dimensions toward a twenty-first-century library catalog | antelman, lynema, and pace examples dynamically displayed based on the index selected. authority searching was relabeled “begins with” searching to let users know that this particular search box featured known-item searching (although it is also where lcsh searching is found) (see figure ). an advanced search tab re-creates the pre-coordinated search options from the web search interface using endeca search functionality. one unique new feature allows users to include or exclude reference materials and government documents from their results. a true boolean search box is made available here, primarily for staff. browse while users can submit a blank search and browse the entire collection by any of the dimensions, the browse tab specifically supports browsing by lc classification scheme (see figure ). this tab also includes a “new titles” browse that can easily be refined with faceted navigation. at the time of this writing, there are plans to pull out other dimensions, such as format, language, or library, for browsing. this will be a great stride forward since there has traditionally been no way to perform a marc codes-only search (in order to browse all chinese fiction in the main library, for example). assessment the endeca-powered catalog seems self-evidently a better tool to help users find relevant resources quickly and intui- tively. but since so much of the implementation involved uncharted territory, plans for assessment began before the launch of the interface, and the actual assessment activi- ties began shortly thereafter. the library identified five assessment measures prior to implementation. one of these, however, requires longer time-series data (changes in circulation patterns), and another, the application of new and potentially complex log-analysis techniques (path analysis). other measures relate to use of the refinements, “sideways searching,” and objective and subjective mea- surements of quality search results, some of which can be preliminarily reported on here. log analysis to learn more about how patrons are using the catalog, data from two months of search logs were analyzed. while authority searching using the library’s old web catalog is still available in the new interface, search logs show that authority searching has decreased percent and keyword searches have increased percent. it is noted, however, that a significant—and indefinable—component of this increase in keyword searching is due to the fact that the default catalog search was changed from title to keyword. users are taking advantage of the new navigational features. fifty-five percent of the endeca-based search requests are simple keyword searches, percent represent searches where users are selecting post-search refinements from the dimensions on the results page, and the remaining percent are true browses with no search term entered (this figure includes use of browse new titles). dimensions the horizontal space just above the results is used to dis- play the full range of results within the lc classification scheme (see figure ). the first dimensions in the left col- umn focus on the subject dimensions (topic and genre) that should be pertinent to the broadest range of searches. the following format and library dimensions recognize that patrons are often limited by time and space. when design- ing the user interface, it was not known which dimensions would be most valuable. as it turned out, dimension use does not exactly parallel dimension placement. lc classification is the most heavily used, followed closely by subject: topic, and then library, format, author, and subject: genre. since no basis for the placement of dimen- figure . new catalog search interface figure . browse by lc classification and new titles information technology and libraries | september sions existed at the time of implementation, the endeca product team plans to use these data, after some time, to determine if changes in dimension order are warranted. spell correction and “did you mean . . .” approximately percent of endeca keyword searches responded to the user’s query with some type of spelling correction or suggestion: . percent performed an auto- matic spell correction, and . percent offered a “did you mean…” suggestion. while ncsu has not analyzed how many of the spell corrections are accurate or how many of the “did you mean…” suggestions are being selected by users, future work in this area is planned. recommender features two features in endeca that have seen a surprising amount of use are the “most popular” sort option and the “more titles like this” feature available on the detailed-record page for a specific title. both relate broadly to the area of recommending related materials to patrons. the “most popular” sort option is currently powered by aggregated circulation data for all items associated with a title. while this technique is ineffective for serials, reference materials, and other noncirculating items, it provides users a previously unavailable opportunity to define relevance. to date, the “most popular” sort is the second most frequently selected sort option (after publica- tion date, at percent), garnering percent of all sorting activity. most-popular sorting was trailed by title, author, and call-number sorting. when viewing a detailed record, users are given the option to find “more titles like this” or “more by these authors.” the first option initiates a new subject keyword search combining the phrases from the $a subdivision of all the subject ( xx) fields assigned to the record. the lat- ter option initiates an author keyword search for any of the authors assigned to the current record. while there are not good statistics on use of this feature, these subject strings appear regularly in the list of most popular queries in search logs. assessing top results if relevance ranking was effective, one would expect to see good results on the first page. but what are “good” or “relevant” results? greisdorf finds that topicality is the first condition of relevance, and xu and chen’s more recent study finds topicality and novelty to be equally important components of relevance. while someone other than the searcher might be able to assess topical relevance, it is impossible to assess novelty, since it cannot be known what the searcher already knows. although researchers agree that relevance is subjec- tive—that is, only a searcher can determine whether results are relevant—janes showed that trained external searchers do a reasonably good job of approximating the topical relevancy judgments of users. the analysis reported here focuses on topicality (using a liberal inter- pretation of what might be topically relevant). ncsu libraries sought to measure how many of the top search results are likely to be relevant to the user ’s query in the old and new catalogs. methodology one of the authors searched topical queries (taken from search logs) in both web and endeca catalogs using “keyword anywhere.” topical queries whose meaning was unclear (e.g., “hand wrought”) were excluded. the topical relevance of the top hits (up to five) was coded for each target. because not all search-result sets contained five records, success for each was measured as a ratio (e.g., / = . ). those searches that resulted in records in both targets were discarded, while those that resulted in records in target a but “found relevant results” in target b were counted as in target a. the ratios were then averaged for each target and compared to determine the difference in relevance-ranking performance. finally, a random subset of forty-four of the queries was selected, and the placement in the web results of the first result in endeca was noted. results on average, percent of the top results in web were judged to be relevant, while percent of the top results in endeca were judged to be relevant. that represents a percent better performance for the endeca catalog. if one makes the assumption that the first endeca record is relevant (admittedly an assumption), based on these data, then one can look at the average position of that record in the old catalog. it was found that the first hit in endeca fell between # and # in web , with more than a third falling after the second screen of results, the maximum number of screens users are typically willing to examine. while this level of increased performance is impres- sive, it masks some dramatic differences in the respec- tive result sets. looking at a broad search, “marsupial,” all of the top five hits in endeca have “marsupial” in the title and “marsupials” or “marsupialia” as a subject heading. the result set includes seventy-eight records, thanks to this intelligent stemming. in the web result set, just twenty-nine records, not a single one of the top five has “marsupial” in the title or subject headings (and the top two results, tributes to malcolm c mckenna and poisonous plants and related toxins, are highly unlikely to be relevant). it is not until record # that you see the first item that contains “marsupial” in the title or subject. this single example demonstrates the benefit of both relevance ranking and stemming. toward a twenty-first-century library catalog | antelman, lynema, and pace usability testing as a result of a long history of catalog-usability studies, there are things that are known about library catalog users. one is that people both expect systems to be easy to use and find that they are not. usability testing was conducted to compare student success in using the new catalog interface with that of students using the old catalog interface when completing the same set of ten tasks. ten undergraduate students were recruited for the test. five were randomly selected to use the old web catalog, while the other five used the new catalog interface, which allows users to choose between a keyword search box powered by endeca and an author- ity search box (begins with . . . ) that is still powered by web . the test contained four known-item tasks and six topical-searching tasks (appendix a). task success, duration, and difficulty were recorded. user satisfaction was not measured since catalog usability studies have found that satisfaction does not correlate with success. task duration figure shows the average task duration for the topical tasks ( – ) for web and endeca. except for task *, there is clearly a trend of significantly decreased average task duration for endeca catalog users. the endeca catalog shows a percent improvement in the average time required to complete a task ( : in web compared to : in endeca). it is also noted that, although results from known-item searching tasks ( – ) are not reported in detail here, test subjects were just as successful in completing them using keyword searching in the endeca catalog as they were using authority searching in web . task success and difficulty in addition to task duration, the test moderator assigned a difficulty rating to each task attempted by the partici- pants: easy, medium, hard, or failed. figure illustrates the overall task-attempt difficulty for topical tasks ( – ) in the web and endeca catalogs. the largest improvement is in the increased percentage of tasks that are completed easily in endeca and the nearly equivalent decrease in the percentage of tasks that were rated as hard to complete. while a significant number of tasks were still failed using the endeca catalog, many of these failures can be attributed to participants’ propensity to select keyword in subject rather than keyword anywhere searches. in fact, the only instances where keyword anywhere search in the new catalog failed to lead to successful task completion were for a single participant who was unwilling to examine retrieved results closely enough to determine if they were actually relevant to the task question, assuming too quickly that the task had been completed successfully. terminology participants using both the web and endeca catalog interfaces expressed confusion over some of the terminol- ogy employed. one of the most problematic terms was “subject.” a number of participants selected keyword in subject for topical searches because of the attraction of the word “subject.” none of the participants recognized that this term referred to controlled vocabulary assigned to records. coupled with a slight unfamiliarity with the term “keyword,” not typically used in web searching, this misunderstanding led participants to misuse (or overuse) keyword in subject searches when they could have found results more effectively using general keyword searching. this terminology problem appears to be an artifact of the usability testing, however. looking at the search logs, more than percent of the keyword searches were keyword anywhere searches, while only percent represented keyword in subject searches. relevance relevance ranking of search results is clearly the most important im-provement in the new catalog. students in this usability test all looked immediately at the first few results on the first page to determine if their search had pro- duced good results. if they didn’t like what they saw, they were likely to retry the search with fewer or more keywords in order to improve their first few results. one participant figure . average task duration: web versus endeca * while task may appear to be an aberration, it actually reveals effec- tive use of new functionality. this task required users to locate an audio recording of poetry in spanish. in web , three of five participants com- pleted the task successfully, all using the material type and language limits available in the advanced search tab. the two participants who didn’t locate this tool failed to complete the task. in endeca, two participants used the same advanced search limits to complete the task success- fully and two additional participants were able to locate and use endeca dimensions to complete the task successfully. this suggests that the new interface is providing users with more options to help them arrive at the results they seek. information technology and libraries | september using the web catalog expressed the need for relevance ranking, “once i scroll through a page, i get pretty discouraged about the results.” the number of pag- ing requests recorded in system logs confirms that users are focusing on the first result screen (with ten results per page); only percent of searchers go to the second page. use of dimensions when questioned after the test, all five par- ticipants who used the endeca catalog intuitively understood that dimen- sions could be used to narrow results. however, only three used the dimensions during the test. throughout the tests, the student participants frequently attempted to limit their search at the outset, rather than beginning with a broad search and then refining. it is unclear whether this behavior is a function of the very specific nature of the test questions or experience with the old catalog. log data show that users are indeed entering broad keyword searches with only one or two terms, which implies that dimensions may be more useful than this usability test indicates. it is also interesting to note that while none of the students understood that the lc classification dimension represented call-number ranges, they did understand that the values could be used to learn about a topic from different aspects—science, medicine, education. ฀ future directions weeks before the initial application went live in january , the list of desired features had grown long. some of these were small “to do” items that the team did not have time to implement. others required deeper investigation, discussion, and testing before the feature could be put into production. still others may or may not be possible. a few of ncsu’s significant planned development directions are summarized below. functional requirements for bibliographic records there is much interest in the utility of applying the functional requirements for bibliographic records model to online catalogs. endeca includes a feature called “record rollup” that allows retailers to group items together for example, different sizes and colors of a shirt. all that is required for this feature is a rollup key. ncsu, working with oclc, has elected to try the oclc work identifier to take advantage of this functionality and create work-level record displays in the endeca catalog hit list. subject access the collective investment libraries have made in subject and name authorities is leveraged with the faceted naviga- tion features of endeca. but only authorized headings in records are seen by endeca, cross-references in the subject- authority record are not used. during implementation, the team looked at ways to improve the entry vocabulary to authorized-subject terms by loading the xx and xx fields from the subject-authority file into endeca synonym tables so that users could be guided to proper subject terms. the team still views this as a promising direction, but simply did not have time to fully explore it prior to implementation. additional discussions with oclc centered on their faceted access to subject terms (fast) project. fast terms are more amenable than lcsh headings to being broken up into topical, geographic, and time-period facets without losing context and meaning. the normalization of geographic and time-period subdivisions promises to be particularly useful. fast has, to date, lacked a ready inter- face for the application of its data. while the fast structure is more conducive to non-cataloger metadata creation and post-coordinate refinement, it still does not meet the need figure . topical task success and difficulty: web versus endeca toward a twenty-first-century library catalog | antelman, lynema, and pace for a user-entry vocabulary. were such a vocabulary for lcsh to become available, it could be mapped to synonym tables to lead users to authorized headings. abandon authority searching? the future of authority searching, however, is less clear. although the usability testing described in this paper showed that the endeca keyword search tools performed on a par with the old catalog for known-item searching, it is recognized that authority searching serves more func- tions. clearly, collocation of all books on a topic is absent when a user does a topical search using keyword rather than a controlled subject heading. but there are more subtle losses as well. as chan points out, one purpose of subject access is to help users focus searches, develop alternative strategies, and enable recall and precision. this is not possible with a simple keyword search, unless the searcher discovers that he can search on a subject heading from a record of interest. the display of subject facets in the endeca-powered catalog works to counter this weakness of simple keyword searching. another navigation aid in the traditional authority dis- play that is lost in a simple keyword-search result is visible “seams.” as mann points out, “seams serve as perceptible boundaries that provide points of reference; without such boundaries, readers get ‘lost at sea’ and don’t know where they are in relation to anything else: they can’t perceive either the extent of what they have, or what they don’t have.” until users have confidence that a known item will appear at the top of a results list if the library holds that item, with a large keyword result set, one cannot confirm a “negative result” without browsing through the entire set. the endeca- powered catalog interface does not help to address either the “seams” or the negative-result problem, which are two reasons why ncsu maintained authority searching. an integration platform despite the vast improvements found in the endeca catalog, the fact remains that it is still mainly books—as calhoun says, “only a small portion of the expanding universe of scholarly information.” there are two approaches to take with the endeca platform: one is to take advantage of having control over the data and the interface to facilitate incorporation of outside data sources to enhance bibliographic records. the second is to put other, non-catalog data sources under the endeca search-and-navigation umbrella. the middleware nature of the endeca platform makes either approach more promising than the “square peg and round hole” problem of trying to work with library management systems ill- equipped to handle a diversity of digital assets. whether as a feed of catalog data to a metasearch application or web-site search tool, or as a platform for faceted access to electronic theses, institutional repositories, or electronic books, endeca has clear potential as a future platform for library resource discovery. ฀ conclusion while it cannot be claimed that this endeca-powered cata- log is a third-generation online catalog, it does implement a majority of the third-generation catalog features identified by hildreth. most notably, through navigation of subject and item-level facets, the endeca catalog supports two of his objectives, “related record search and browse” and “integration of keyword, controlled vocabulary, and clas- sification-based approaches.” spell correction, intelligent stemming, and synonym tables support “automatic term conversion/matching aids.” the flexible relevance-rank- ing tools support “closest, best-match retrieval” as well as “ranked output.” much work remains, however. three important features identified by hildreth cannot be said to be implemented in this catalog at this time: “natural language query expression,” that is, an entry vocabulary, “expanded coverage and scope,” and “relevance feedback methods.” requirements for these features are either being reviewed or are already under development by both endeca and ncsu libraries. ncsu views the endeca catalog implementation in the context of a broader, critical evaluation and overhaul of library discovery tools. like the library web site, the catalog still requires users to come to it. when they do, it still sets a high threshold for patience and the ability to interpret clues. still, at the end of the day it rewards the ncsu student searching “declaration of independence” with the book, american scripture: making the declaration of independence instead of the recent congressional resolution, recognizing the mexican holiday of cinco de mayo. references . christine l. borgman, “why are online catalogs still hard to use?” journal of the american society for information sci- ence , no. ( ). . karl v. fast and d. grant campbell, “i still like google: university student perceptions of searching opacs and the web.” in proceedings of the th asis&t annual meeting (providence, r.i.: american society for information science and technology, ). . ray r. larson, “between scylla and charybdis: subject searching in the online catalog,” advances in librarianship ( ); andrew large and jamshid beheshti, “opacs: a research review,” library & information science research , no. ( ). . nathalie nadia mitev, gillian m. venner, and stephen walker, designing an online public access catalogue: okapi, a cata- logue on a local area network (london: british library, ). information technology and libraries | september . borgman, “why are online catalogs still hard to use?” . . r. hafter, “the performance of card catalogs: a review of research,” library research , no. ( ). . gerard salton, “the use of extended boolean logic in information retrieval,” in proceedings of the acm sigmod international conference on management of data (new york: acm pr., ), . . ray r. larson, “classification clustering, probabalistic information retrieval, and the online catalog,” library quar- terly , no. ( ). . ibid. . charles r. hildreth, online public access catalogs: the user interface (dublin, ohio: oclc, ). . larson, “classification clustering.” . mitev, venner, and walker, designing an online public access catalogue; ray r. larson et al., “cheshire ii: designing a next-generation online catalog,” journal of the american society for information science , no. ( ). . tamas e. doszkocs, “cite nlm: natural-language search- ing in an online catalog,” information technology and libraries , no. ( ). . nicholos j. belkin and w. bruce croft, “retrieval tech- niques,” in annual review of information science and technology, ed. martha e. williams (new york: elsevier, ), . . gary marchionini, information seeking in electronic envi- ronments (new york: cambridge univ. pr., ), – . . borgman, “why are online catalogs still hard to use?” . . lois mai chan, exploiting lcsh, lcc, and ddc to retrieve networked resources: issues and challenges (washington, d.c.: library of congress, ), www.loc.gov/catdir/bibcontrol/ chan_paper.html (accessed july , ). . lois mai chan, “library of congress classification as an online retrieval tool: potentials and limitations,” information technology and libraries , no. ( ); mary micco and rich popp, “improving library subject access (ilsa): a theory of clustering based in classification,” library hi tech , no. ( ). . marcia j. bates, “subject access in online catalogs: a design model,” journal of the american society for information sci- ence , no. ( ); karen coyle, “catalogs, card—and other anachronisms,” the journal of academic librarianship , no. ( ); larson, “classification clustering.” . karen markey, “thus spake the opac user,” information technology and libraries , no. ( ): . . larson, “classification clustering.” . marcia j. bates, library of congress bicentennial conference on bibliographic control for the new millennium, task force recom- mendation . research and design review: improving user access to library catalog and portal information, final report, ; charles r. hildreth, intelligent interfaces and retrieval methods for subject searching in bibliographic retrieval systems (washington, d.c.: library of congress, ); bates, “subject access in online catalogs”; belkin and croft, “retrieval techniques.” . bates, “subject access in online catalogs”; bates, library of congress bicentennial conference on bibliographic control for the new millennium; eric novotny, “i don’t think, i click: a protocol analysis study of use of a library online catalog in the internet age,” college & research libraries , no. ( ). . bates, “subject access in online catalogs,” . . larson, “classification clustering”; buckland et al., “map- ping entry vocabulary to unfamiliar metadata vocabularies,” d-lib magazine , no. ( ). . h. greisdorf, “relevance thresholds: a multi-stage pre- dictive model of how users evaluate information,” information processing & management , no. ( ): – ; yunjie (calvin) xu and zhiwei chen, “relevance judgment: what do informa- tion users consider beyond topicality?” journal of the american society for information science and technology , no. ( ). . joseph w. janes, “other people’s judgments: a compari- son of users’ and others’ judgments of document relevance, topicality, and utility,” journal of the american society for informa- tion science , no. ( ). . bernard j. jansen and udo pooch, “a review of web searching studies and a framework for future research,” jour- nal of the american society for information science and technology , no. ( ); novotny, “i don’t think, i click.” . borgman, “why are online catalogs still hard to use?” . brian nielsen and betsy baker, “educating the online catalog user: a model evaluation study,” library trends , no. ( ). . ifla cataloging section, “frbr bibliography,” www.ifla .org/vii/s /wgfrbr/bibliography.htm (accessed may , ). . lois mai chan et al., “a faceted approach to subject data in the dublin core metadata record,” journal of internet catalog- ing , no. / ( ). . chan, exploiting lcsh, lcc, and ddc. . thomas mann, “is precoordination unnecessary in lcsh? are web sites more important to catalog than books?” a refer- ence librarian’s thoughts on the future of bibliographic control (washington, d.c.: library of congress, ), www.loc.gov/ catdir/bibcontrol/mann_paper.pdf (accessed july , ). . karen calhoun, “the changing nature of the catalog and its integration with other discovery tools,” prepared for the library of congress, , . unpublished, www.loc.gov/ catdir/calhoun-report-final.pdf (accessed july , ). . charles r. hildreth, online catalog design models: are we moving in the right direction? (washington, d.c.: council on library resources, ). toward a twenty-first-century library catalog | antelman, lynema, and pace copyright © by charles w. bailey jr. this work is licensed under the creative commons attribution- noncommercial . license. to view a copy of this license, visit http://creativecommons.org/licenses/by-nc/ . / or send a letter to creative commons, howard st., th floor, san francisco, ca, , usa. bailey continued from ฀ known-item questions . “your history professor has requested you to start your research project by looking up background information in a book titled civilizations of the ancient near east.” a. “please find this title in the library catalog.” b. “where would you go to find this book physically?” . “for your literature class, you need to read the book titled gulliver’s travels written by jonathan swift. find the call number for one copy of this book.” . “you’ve been hearing a lot about the physicist richard feynman, and you’d like to find out whether the library has any of the books that he has written.” a. “what is the title of one of his books?” b. “is there a copy of this book you could check out from d. h. hill library?” . “you have the citation for a journal article about photosynthesis, light, and plant growth. you can read the actual citation for the journal article on this sheet of paper.” alley, h., m. rieger, and j.m. affolter. “effects of developmental light level on photosynthesis and biomass production in echinacea laevigata, a federally listed endan- gered species.” natural areas journal . ( ): – . a. “using the library catalog, can you determine if the library owns this journal?” b. “do library users have access to the volume that actually contains this article (either electronically or in print)?” ฀ topical questions . “please find the titles of two books that have been written about bill gates (not books written by bill gates).” . “your cat is acting like he doesn’t feel well, and you are worried about him. please find two books that provide information specifically on cat health or caring for cats.” . “you have family who are considering a solar house. does the library have any materials about building passive solar homes?” . “can you show me how would you find the most recently published book about nuclear energy policy in the united states?” . “imagine you teach introductory spanish and you want to broaden your students’ horizons by expos- ing them to poetry in spanish. find at least one audio recording of a poet reading his or her work aloud in spanish.” . “you would like to browse the recent journal litera- ture in the field of landscape architecture. does the design library have any journals about landscape architecture?” appendix a: ncsu libraries catalog usability test tasks president's column information technology and libraries | september b eing president of a dynamic organization like lita is truly a humbling experience. every day i am awestruck by the dedication, energy, creativity, and excitement exhibited by lita’s members. i see it in everything that lita does, from its stellar publications and communications—including this journal, ital—to its programming and contribution to standards and system development. none of this would be possible without the hard work of all the dedicated members who volunteer their time not only to advancing their own professional development, but also to advancing the profession. thank you all. for forty years now, lita members have been dedicated to the association’s work, and we have been celebrating our fortieth anniversary throughout . the celebration continues as we prepare to convene in nashville for the ninth lita national forum, october – , . lita has had a long tradition of providing quality conferences. the first, held in , was the conference on interlibrary communications and information networks, more familiarly known as the “airlie conference,” which had published proceedings. the second was a cooperative effort held in with the library education division and the american society for information science (asis), entitled “directions in education for information science: a symposium for educators.” in later years, lita held three national conferences: baltimore ( ), boston ( ), and denver ( ). in , lita and the library administration and management association (lama) held a joint conference in pittsburgh. while the national conferences were very successful, the idea of a more infor- mal, intimate event to be held annually took form, and in lita held its first annual national forum. next year we will continue the tradition of successful conference programming as we celebrate the tenth anniversary of the lita national forum in denver. this year’s theme is “netville in nashville: web services as library services.” we have an exciting lineup of keynote and concurrent-session speakers as well as several poster-session presenters who will stimulate lively discussions in all of the wonderful, informal networking opportunities this small conference offers. the sponsor showcase allows plenty of time for attendees to talk to our valued sponsors and learn more about their products. the two preconference programs offer in-depth experiences: “opensource installfest” and “developing best project management practices for it projects.” lita bloggers will be out in force producing summaries and reactions to it all. one of lita’s strongest membership benefits is the per- sonal networking opportunities it provides. by providing an informal and enjoyable atmosphere, the national forum is one of the best places to network with others dealing with the same issues as you. i hope to see you there. besides the national forum (just one of lita’s many educational programs), one of the things i like most about lita is its flexibility to quickly accommodate program- ming to cover the latest issues and trends. lita’s pro- gramming at ala annual conferences attracts attendees from all divisions for this reason. every year, the highly successful top technology trends attracts more and more people who come to listen to the experts speak on the lat- est trends. the lita interest groups, like the technologies they focus on, also exhibit great flexibility because they can come and go—it’s easy to locate a few other members to create a new group where interested parties can come together for focused discussions or formal presentations. since its inception, lita has had traveling educational programs to provide programming opportunities for people who cannot attend the ala conferences. these in-depth programs, now called the regional institutes, focus on a topic and are offered as long as that issue is relevant. look for new electronic delivery of lita pro- grams in the future. of course, lita’s publications provide a very lasting educational component. lita launched journal of library automation (jola), the predecessor of ital, in , one year after the formation of the new division of ala. jola and, later, ital have consistently been a place for library information technologists to publish in a peer-reviewed scholarly journal. these well-respected publications have had a wonderful group of editors and editorial boards over the years. we are pleased that ital is now available online for members from the moment of publication. i want to thank all the people who work so hard to produce this publication on a quarterly basis. i also want to thank all the authors who submit their research for publication here and make a lasting contribution to the profession. all of these programs are just a sampling of what lita provides its members. is it any wonder i am awed by it all? i hope you are as well. i also hope that, in my year as your president, you will communicate with me in an open dialogue on the lita blog, via e-mail, or in person at conferences regarding how lita can better meet your needs as a member. we have been focusing a great deal on our educational goal because that is what we have heard you want out of lita. i encourage you to let me and the rest of the lita board know how we can best deliver a quality set of educational programs. president’s column bonnie postlethwaite bonnie postlethwaite (postlethwaiteb@umkc.edu) is lita pres- ident / and associate dean of libraries, university of missouri–kansas city. fagan information technology and libraries | september visual search interfaces have been shown by researchers to assist users with information search and retrieval. recently, several major library vendors have added visual search interfaces or functions to their products. for pub- lic service librarians, perhaps the most critical area of interest is the extent to which visual search interfaces and text-based search interfaces support research. this study presents the results of eight full-scale usability tests of both the ebscohost basic search and visual search in the context of a large liberal arts university. l ike the web, online library research database inter- faces continue to evolve. even with the smaller scope of library research databases, users can still suffer from information overload and may have difficulty in processing large results sets. web search-engine research has shown that the number of searchers viewing only the first results page has increased from percent in to percent in for united states-based web search- engines users. additionally, the mean number of results viewed per query in was . documents. this may indicate either increasing relevance in search results or an increase in simplistic web interactions. visual alternatives to search interfaces attempt to address some of the problems of information retrieval within large document sets. while research and devel- opment of visual search interfaces began well before the advent of the web, current research into visual web interfaces has continued to expand. within librarianship, the most visual interface research seems to focus on those that could be applied to large-scale digital library projects. although library products often have more metadata and organizational structure than the web, search engine-style interfaces adapted for field searching and boolean opera- tors are still the most frequent approach to information retrieval. yet research has shown that visual interfaces to digital libraries offer great benefit to the user. zaphiris emphasizes the advantage of shifting the user’s mental load “from slow reading to faster perceptual processes such as visual pattern recognition.” according to borner and chen, visual interfaces can help users better under- stand search results and the interrelation of documents within the result set, and refine their search. in their dis- cussion of the function of “overviews” in visual interfaces, greene and his colleagues say that overviews can help users make better decisions about potential relevance, and “extract gist more accurately and rapidly than traditional hit lists provided by search engines.” several library database vendors are implement- ing visual interfaces to navigate and display search results. serials solutions’ new federated search product, centralsearch, uses technology from vivisimo that “orga- nizes search results into titled folders to build a clear, concise picture for its users.” ulrich’s fiction connection web site has used aquabrowser to help one “discover titles similar to books you already enjoy.” the queens library has also implemented aquabrowser to provide a graphical interface to its entire library’s collections. xreferplus maps search results to topics by making visual connections between terms. comabstracts, from cios, uses a similar concept map, although one cannot launch a search directly from the tool. groxis chose a circular style for its concept-mapping software, grokker. partnerships between groxis and stanford university began as early as , and grokker is now being implemented at stanford university libraries academic and information resources. ebsco and groxis announced their partnership in march . the ebscohost interface now features a visual search tab as an option that librarians can choose to leave on (by default) or turn off in ebsco’s administrator module. figure shows a screenshot of the visual search interface. within the context of library research databases, visual searching likely provides a needed alternative from tradi- tional, text-based searching. to test this hypothesis, james madison university libraries (jmu libraries) decided to conduct eight usability sessions with ebscohost’s new visual search, in coordination with ebsco and groxis. while this is by no means the first published usability test of vendor interfaces, the literature understandably reveals a far greater number of usability tests on in-house projects such as library web sites and customized catalog interfaces than on library database interfaces. it is hoped that by observing users try both the ebsco basic search and visual search, more understanding will be gained about user search behav- ior and the potential benefits of a visual approach. ฀ method the usability sessions were conducted at jmu, a large liberal arts university whose student population is mostly drawn from virginia and the northeastern region. only percent of the students are from minority groups. jmu requires that all freshmen pass the online information skills seeking test (isst) before becoming a sophomore, and the libraries developed a web tutorial, “go for the gold,” to prepare students for the isst. therefore, usabil- jody condit fagan usability testing of a large, multidisciplinary library database: basic search and visual search jody condit fagan (faganjc@jmu.edu) is digital services li- brarian at carrier library, james madison university, har- risonburg, virginia. usability testing of a large, multidisciplinary library database | fagan ity-test participants were largely white, from the northeast- ern united states, and had exposure to basic information literacy instruction. jmu libraries’ usability lab is a small conference room with one computer workstation equipped with morae soft- ware. audio and video recordings of user speech and facial expressions, along with “detailed application and computer system data,” are captured by the software and combined into a searchable recording session for the usability tester to review. a screenshot of the morae analysis tool is shown in figure . the usability test script was developed in collabora- tion with representatives of ebsco and groxis. ebsco provided access to the beta version of visual search for the test, and groxis provided financial incentives for student participants. the test sessions and the results analysis, however, were conducted solely by the researcher and librarian facilitators. the visual search development team was provided with the results and video clips after analysis. usability study participants were recruited by posting an announcement to the jmu students’ web portal. a $ gift certificate was offered as an incentive, and more than students submitted a participation interest form. these were sorted by the number of years the student(s) had been at jmu to try to get as many novice users as possible. because so much of today’s student work is conducted in groups, four groups of two, as well as four individual ses- sions, were scheduled, for a total of twelve students. jmu librarians who had received both human-subjects training and an introduction to facilitation served as facilitators to the usability sessions. their role was to watch the time and ask open-ended questions to keep the student participants talking about what they were doing. the major research question it was hoped would be answered by the tests was, “to what extent does ebsco’s basic search interface and visual search interface support student research?” since the tests could not evaluate the entire research process, it was decided to focus on the development of the research topic. specifically, the goal was to find out how well each interface supported the intellectual process of the students in coming up with a topic, narrowing their topic, and performing searches on their chosen subtopics. an additional goal was to deter- mine how well users were able to find and use the interface widgets and how satisfied the students felt after using the interfaces. the overall session was structured in this order: a pretest survey about the students’ research experience; a series of four tasks performed with ebscohost’s basic search; a series of three tasks performed with ebscohost’s visual search; and a posttest interview. both basic and visual search interfaces were used with academic search premier. each of the eight sessions was recorded in entirety by the morae software, and each recording was viewed in entirety. to try to gain some quantitative data, the researcher measured the time it took to complete each task. however, due to variables such as facilitator involvement and interaction between group members, the numbers did not lend themselves to comparison. also, it would not have been clear whether greater numbers indicated a positive or negative sign. taking longer to come up with subtopics, for example, could as easily be a sign of exploration and interested inquiry as it might be of frustration or failure. as such, the data are mostly qualitative in nature. figure . screenshot of ebscohost’s visual search figure . screenshot of morae recorder analysis tool information technology and libraries | september ฀ results the student participants were generally underclassmen. two of the students, group , were in their third year at jmu. all others were in their first or second year. while students were drawn from a wide variety of majors, it is regrettable that there was not stronger representation from the humanities. when asked, “what do you normally use to do research?” six students answered an unquali- fied “google.” three other students mentioned internet search engines in their response. only two students gave the brand or product names of library research databases: one said, “pubmed, wilsonomnifile, and ebsco,” while the other, a counseling major, mentioned psycinfo and cinahl. when shown a screenshot of basic search, half of the students said they had used an ebsco database before. all of the participants said they had never before used a visual search interface. the full results from the individual pretest interviews are shown in figures and . to begin the usability test, the facilita- tor started internet explorer and loaded the ebscohost basic search, which was set to have a single input box. the scripts for each task are listed in figure . note that task was only featured in the basic search portion of the test. for task on the basic search—coming up with a general topic—all of the par- ticipants began by using their own topics rather than choosing from the list of ideas. also, although they were asked to “spend some time on ebsco to come up with a possible general topic,” all but group fulfilled this by simply thinking of a topic (sometimes after some discussion within the groups of two) and typing it in. with the exception of group , the size of the result set did not inspire topic changes. figure summarizes the students’ searches and relative success on task . in retrospect, the tests might have yielded more straightforward findings if the students had been directed to choose from the provided list of topics, or even to use the same topic. however, part of the intention was to determine whether either interface was helpful in guiding the students’ topic development. it was hoped that by defining the scenario as writing a paper for class, their topic selection would reflect the realities of student research. however, it probably would have been better to have used the same topic for each session. task asked participants to identify three subtopics, and task asked them to refine their search to one subtopic and limit it to the past two years. a summary of these tasks appears in figure . a surprising finding during task was that students did go past the first page of results. four groups went past the first page of results, while two groups did not get enough results for more than one page. the other two groups did not choose to look past the first page of results. this contrasts with jansen and spink’s findings, figure . results from pretest interview, groups – figure . results from pretest interview, groups – usability testing of a large, multidisciplinary library database | fagan in which percent of web searchers only view the first results page. another pleasant surprise was that students spent some time actually reading through results when they were searching for ways to narrow their topic. five groups scanned through both titles and abstracts, which requires clicking on the article titles to display the citation view. one of these five additionally chose to open full-text articles and look at the references to determine relevance. two groups scanned through the results pages only, but looked at both article titles and the subjects in the left-hand column. group seemed to only scan the titles in the results list. this user behavior is also quite different than that found with web search-engine users. in one recent study by jansen and spink, more than percent of the time, search-engine users viewed five or fewer documents per query. the five groups that chose to view the citation/abstract view by clicking on the title (groups , , , , and ) identi- fied subtopics that were significantly more interesting and plausible than the general topic they had come up with. from looking at their results, these groups were clearly identifying their subtopics from reading the abstracts and titles rather than just brainstorming. although group had the weakest subtop- ics, going from the world baseball classic to specific players’ relationships to the classic and the home-run derby, they were working with a results set of but eleven items. the three groups that relied on scanning only the results list succeeded to an extent, but as a whole, the new subtopics would be much less satisfying to the scenario’s hypo- thetical professor. after scanning the titles on two pages of results, group (an individual) ended up brainstorming her subtopics (pre- vention, intervention, and what an eating disorder looks like) based on her knowledge of the topic rather than drawing from the results. group (a group of two) identified their subtopic (sand dunes) from the left- hand column on the results list. group (an individual) picked up his subtopics (steroids in sports, president bush’s stance on steroids, and softball) from reading keywords in the article titles on the first page of results. since the subjects in the left-hand column were a new addition to basic search, the use of this area was also noted. four groups used the subjects in the left-hand column without prompting. two groups saw the subjects (i.e., ran the mouse over them) but did not use them. the remaining two groups made no action related to the subjects. a worrisome finding of tasks and was that most students had trouble with the default search being set to phrase-searching rather than to a boolean and. this can easily be seen in looking at the number of results the students came up with when they tried to refine their topics (figure ). even though most students had some limiter still in effect (full text, last two years) when they first tried their new refined search, it was the phrase- searching that really hurt them. luckily, this figure . task , coming up with a general topic using basic search figure . tasks posed for each portion of the usability test. information technology and libraries | september is a customizable setting in ebsco’s administrator module, and it is recommended that libraries enable the “proximity” expander to be set “on” by default, which will automatically combine search terms with and. task , finding a “recent article in the economist about the october earthquake in kashmir,” was designed to test the usability of the ebscohost publication search and limiter. it was listed as optional in case the facilitator was worried that time was an issue. four of the student groups— , , , and —were posed the task. of these four groups, three relied entirely on the publication lim- iter on the refine search panel. group chose to use the publication search. all four groups quickly and success- fully completed this task. ฀ ฀additional questions during basic search tasks at various points during the three tasks in ebsco’s basic search, the students were asked to limit their results set to only full-text results, to find one peer-reviewed article, and to limit their search to the past two years. seven out of the eight student groups had no problem finding and using the ebscohost “refine search” panel, including the full-text check box, date limiter, and peer- reviewed limiter. group did not find the refine search panel or use its limiters until specifically guided by the facilitator near the end. this group had found other ways to apply limits: they used the “books/monographs” tab on the results list to limit to full text, and the results-list sorting function to limit to the past two years. after having seen the refine search panel, group did use the “peer reviewed” check box to find their peer-reviewed article. toward the end of the basic search portion, students were asked to “save three of their results for later.” three groups demonstrated full use of the folder. an additional three groups started to use the folder and viewed the folder but did not use print, save, or e-mail. it is unclear whether they knew how to do so and just did not follow through, or whether they thought they had safely stored the items. two students did not use the folder at all, act- ing individually on items. one group used the “save” function but did not save each article. ฀ visual search similar to task , when using the basic search, students did not discover general topics by using the interface, but simply typed in a topic of interest. only two groups, and , chose to try the same topic again. in the interests of processing time, visual search limits the search to the first results retrieved. since jmu has set the default sort results to display in chronological order, the most recent results were returned during these usability tests. figure shows the students’ original search terms using visual search, the actions they took while looking for subtopics, and the subtopics they identified. additionally, if the subtopics they identified matched words on the screen, the location of those words is noted. three of the groups ( , , and ) identified subtopics when looking at the labels on topic and subtopic circles. group identified subtopics while looking at article titles as well as the subtopic circles. the members of group identified subtopics while look- ing at the citation view and reading the abstract and full text, as well as rolling over article titles with their mice. it was not entirely clear where the student in group got his subtopics from. two of the three subtopics did not seem to figure . basic search, task and , coming up with subtopics. usability testing of a large, multidisciplinary library database | fagan be represented in the display of the results set. his third subtopic was one of the labels from a subtopic circle. groups and both struggled with finding their subtopics. group simply had a narrow topic (“jackalope”), and group misspelled “steroids” and got few results for that reason. lacking many clusters, both groups tried typing additional terms into the title keyword box on the filter panel, resulting in fewer or zero results. for task , students were asked to limit their search to the last two years and to refine their search to a chosen subtopic (figure ). particularly because the results set is limited to , it would have been better to have separated these two tasks: first to have them limit the content, then perhaps the date of the search. three groups, all groups of two, used the date limit first ( , , and ). three groups ( , , and ) narrowed the content of their search by typing a new search or additional keywords into the main search box. groups and narrowed the content of their search by clicking on the subtopic circles. note that this does not change the count of the number of results displayed in the filter panel. groups and tried typing keywords into the title keyword filter panel and also clicking on circles. both groups fared better with the latter approach. group typed an additional keyword into the filter panel box to narrow his search. while five of the groups announced the subtopic to which they wanted to narrow their search before beginning to narrow their topic, groups , , and began to interact with the interface and experi- ment with subtopics before choosing one. while groups and arrived at a subtopic and identified it, group tried many experiments, but since their original topic (jackalope) was already narrow, they were not ultimately successful in identifying or searching on a subtopic. as with basic search, students were asked to save three articles for later. five of the groups ( , , , , and ) used the “add to folder” function which appears in the citation view on the right-hand side of the screen. of these, three groups proceeded to “folder has items.” of these groups, two chose the “save” function. two groups used either “save” or “e-mail” to preserve individual items, rather than using the folder. one group experienced system slowness and was not able to load the full-record view in time to deter- mine whether they would be able to save items for later. a concern that students may not realize is that in folder view or individually, the “save” button really just formats the records. the user must still use a browser function to save the formatted page. no student performed this function. figure . visual search, task and , coming up with a general topic figure . visual search, task , searching on subtopic (before date limit, if possible) information technology and libraries | september several students had some trouble with the mechanics of the filter panel, shown in figure . seven of the eight groups found and used the filter panel, originally hidden from view, without assistance. however, some users were not sure how the title keyword box related to the main search box. at least two groups typed the same search string into the title keyword box that they had already entered into the main search box. also, users were not sure whether they needed to click the search button after using the date limiter. however, in no case was a student unable to quickly recover from these areas of confusion. ฀ results of posttest interview at the end of the entire usability session, participants were asked several questions while looking at screenshots of each interface. a full list of posttest interview questions can be found in figure . when speaking about the strengths of basic search, seven of eight groups talked about the search options, such as field searching and limiters. the individual in group mentioned “the ability to search in fields, especially for publications and within publications.” one of the students in group mentioned that “i thought it was easier to specify the search for the full text and the peer reviewed—it had a separate page for that.” the student in group added, “they give you all the filter options as opposed to the other one.” five of the eight groups also mentioned familiarity with the type of interface as a strength of basic search. since jmu has only had access to ebsco databases for less than a year, and half of the students admitted they had not used ebsco, it seemed their com- ments were with the style of interface more than their experience with the interface. the student in group commented, “seems like the standard search engine.” group noted, “it was organized in a way that we’re used to more,” and group said, “it’s more traditional so it’s more similar to other programs.” half of the groups mentioned that basic search was clear or organized. group explained, “it was nice how it was really clearly set out . . . like, everything’s in a line.” not surprisingly, visual search’s strengths surrounded the grouping of subtopics: seven of eight groups made some comment about this. the student in group said, “it groups the articles for you better. it kinda like gives you the subtop- ics when you get into it and search it and that’s pretty cool.” the student in group stated, “you can look and see an outline of where you want to go . . . it’s easy to pinpoint it on screen like that’s where i want to go with my research.” some of the other strengths mentioned about visual search were: showing a lot of information on one screen without scrolling (group ) and the colorful nature of the interface. a student in group added, “i like the circles and squares—the symbols register easily.” the only three weaknesses listed for basic search in response to the first question were: “not having a spot to put in words not to search for” (group ); that, like internet search engines, basic search should have “a clip from the article that has the keyword in it, the line before and the line after” (group ); and that basic search might be too broad, because “unless you narrow it, [you have to] type in keywords to narrow it down yourself” (group ). figure . visual search filter panel figure . posttest interview questions usability testing of a large, multidisciplinary library database | fagan with regard to weaknesses of visual search, half of the groups had some confusion about the content, partially due to the limited number of results. a student from group declared, “it may not have as many results. . . . if you typed in ‘school’ on the other one, it might have . . . , pages [but] on this you have . . . results.” the student in group agreed, saying that with visual search, “they only show you a certain number of articles.” the student in group said, “it’s kind of confusing when it breaks it up into the topics for you. it may be helpful for some other people, but for the way my mind works i like just having all my results displayed out like on the regular one.” half of the groups also made some comment that they were just not used to it. six of the groups were asked which one they would choose if they had class in one hour. (it is not clear why the facilitator did not ask this question of groups and .) four groups ( , , , and ) indicated basic search. one student in group said, “i think it’s easier to use, but i don’t trust it.” the other in group added, “it’s new and we’re not quite sure because every other search engine is you just type in words and it’s not graphical.” both stu- dents in group commented that the familiarity of basic search was the reason they would use it for class in one hour. both groups and would later say that they liked the visual search interface better. two groups ( and ) chose visual search for the “class in one hour” scenario. the student in group commented, “because it does cool things for you, makes it easier to find. otherwise you’re going through by title.” both these groups would later also say that they liked the visual search interface better. the students were also asked to describe two scenarios, one in which they would use basic search and one in which they would use visual search. four of the groups ( , , , and ) said they would use basic search when they knew what information they needed. seven of the eight groups said they would use visual search for broad topics. all the students’ responses are given in figure . when asked which interface they preferred, the groups split evenly. comments from the four who preferred basic search ( , , , and ) centered on the familiarity of the interface. the student in group added, “the regular one . . . i like to get things done.” all four of these students had said they had used an ebsco database before. the two students who could list library research databases by name were both in this group. of the four who preferred visual search ( , , , and ), three groups had never used ebsco before, though one of the students in group thought he’d used it in the library web tutorial. group commented, “it seemed like it had a lot more information . . . cool . . . futuristic.” the student in group said, “it’s kind of like a little game. . . . like you’re trying to find the hidden piece.” group commented that visual search was colorful and intriguing. the students in group both stated “the visual one” in unison. one stu- dent said that visual search was more “[eye-catching] . . . it keeps you focused at what you are doing, i felt, instead of . . . words . . . you get to look at colors” and added later that it was “fun.” the other students in group said, “i’m a very visual learner. so to see instead of having to read the categories, and say oh this is what makes sense, i see the circles like ‘abilities test’ or ‘academic achievement’ and i automatically know that’s what it is . . . and i can see how many articles are in it . . . and you click on it and it zooms in and you have all of them there.” the second student went on to add, “i’ve been teaching my mom how to use technology and the visual search would be so much easier for her to get, because its just looks like someone drew it on there like this is a general category and then it breaks it down.” other suggestions given during the free-comment portion of the survey were to have the filters from basic search appear on visual search (especially peer-reviewed); curiosity about when visual search would become available (at the time it was in beta test); and a suggestion to have general- education writing students write their first paper using visual search. figure . examples of two situations: one in which you would be more likely to use visual search, and one in which you would be more likely to use ebsco information technology and libraries | september ฀ discussion this evaluation is limited both because most students chose different topics for each search interface, and because they only had time to research one topic in each interface. therefore, there could be an infinite number of scenarios in which they would have performed differently. however, this study does show that, for some students, or for some search topics, visual search will help students in a way that basic search may not. one hypothesis of this study was that within the con- text of library research databases, visual searching would provide a needed alternative from traditional, text-based searching. the success of the students was observed in three areas: the quality of the subtopics they identified after interacting with their search results; the improvement of the chosen subtopic over their chosen general topic, and the quality of the results they found for their subtopic search. the researcher made a best effort to compare top- ics and results sets and decide which interface helped the student groups to perform better. in addition, qualities that each interface seemed to contribute to the students’ search process were noted (figure ). these qualities were deter- mined by reviewing the video recordings and examining the ways in which either interface seemed to support the attitudes and behaviors of the students as they conducted their research tasks. when considering all three of these areas, four groups did not, overall, require visual search as an alternative to basic search ( , , , and ). two of these groups ( and ) seemed to benefit from more focus when using the basic search interface. although visual search lent them more interaction and exploration (which may be why they said they preferred visual search), it seems the focus was more important to their performance. for the other two groups ( and ), basic search really supported the depth of inquiry and high interest in find- ing results. these two groups confirmed that they preferred basic search. for two groups ( and ), visual search seemed an equally viable alternative to basic search. for group , both interfaces seemed to support the group’s desire to explore; they said they preferred visual search. for the student in group , basic search seemed to orient him to the goal of finding results, while visual search sup- ported a more exploratory approach. since, in his case, this exploratory approach did not turn out well in the area of finding results, it is not surprising that he ended up preferring basic search. the remaining two groups ( and ) performed bet- ter with visual search, upholding the hypothesis that an alternate search is needed. group seemed bored and uninterested in the search process when using basic search even though they chose a topic of personal interest: “world baseball classic.” visual search caught their attention and sparked interest in the impersonal topic “global warming.” group spent more time exploring while using the visual search interface, and in the posttest survey admitted that they preferred the visual search interface. the student in group said she preferred basic search, and as a self- described psycinfo user, seemed comfortable with the interface. yet for this test scenario, visual search made her think of new ideas and supported more real exploration during the search process. within each of the three areas, basic search appeared to have the upper hand for both the quality of the subtopics identified by the students, and in the improvement of the chosen subtopics over the general topics. this is at least partially explained by the limitation of visual search to the most recent results. that is, as the students explored the visual search results, choosing subtopics would not relaunch a search on that subtopic, which would have engendered more and perhaps better subtopics. in the third area, the quality of the results set for the chosen topic, visual search seemed to have the upper hand if only because of the phrase-searching limitation present in jmu’s administrative settings for basic search. that is, students were often finding few or no results on their chosen subtopics in basic search. this study also had findings that seem to transcend figure : strengths of basic search and visual search in quality of subtopics, most improved topic, and result sets usability testing of a large, multidisciplinary library database | fagan these interfaces and the underlying database. first, librar- ies should strongly consider changing their database default searching from phrase searching to a boolean and, if possible. (this is possible in ebsco using the administrative module.) second, most students did not have trouble finding or using the interface widgets to perform limiting functions, with the one exception being some confusion about the relationship between the visual search filters and main search box. unlike some research into web search behavior, students may well travel beyond the first page of results and view more than just a few docu- ments when determining relevance. finally, the presence of subject terms in both interfaces proved to be an aid to understanding results sets. this study also pointed out some improvements that could be made to visual search. first, it would be great if visual search returned more than results in the initial set, or at least provided an overview of the size, type, and extent of objects using available metadata. however, even with today’s high-speed connections, result-set size will need to be balanced with performance. perhaps, as students click on subtopics, the software could rerun the search so that the results set does not stay limited to the original . on a minor note, for both basic and visual search, greater care should be taken to make sure users understand how the save function works and alert users to the need to use the browser function to complete the process. it should be noted that ebsco has not stopped devel- oping visual search, and many of these improvements may well be on their way. ebsco says it will be adding more support for limiters, display preferences, and contextual text result-list viewing at some point in the future. these feature sets can currently be viewed on grokker.com. an important area for future research is user behavior in library subscription databases. while these usability tests provide a qualitative evaluation of a specific inter- face, it would be worthwhile to have a more reliable understanding about students’ searching behavior in library databases across similar interfaces. since public service librarians deal primarily with users who have self-identified as needing help, their experience does not always describe the behavior of all users. furthermore, studies of web search behavior may not apply directly to searching in research databases. specifically, students’ use of subject terms in both interfaces could be explored. half of the student groups in this study chose to use the basic search subject clusters in the left-hand column on the results page, despite the fact that they had never seen them before (this was a beta-test feature). is this typical? would this strategy hold up to a variety of research topics? another interesting question is the use of a single search box versus several search boxes arrayed in rows (to assist in constructing boolean and field searching). in the ebsco administrative module, librarians can choose either option. based on research rather than anecdotal evidence, which is best? another option is the default sort: historically, at jmu libraries, this has been a chronological sort. does this cause problems for relevance-thinking students? finally, the issue of collaboration in student research using library research databases would be a fascinat- ing topic. certainly, these usability recordings could be reviewed with a mind to capturing the differences between individuals and groups of two, but there may be better designs for a more focused study of this topic. ฀ conclusion if you take away one conclusion from this study, let it be this: do not hesitate to try visual search with your users! information providers must balance investments in cut- ting-edge technology with the demands of their users. libraries and librarians, of course, are a key user group for information providers. a critical need in librarianship is to become familiar with the newest technology solutions, particularly with regard to searching, in order to provide vendors with informed feedback about which technologies to pursue. by using and teaching new visual search alter- natives, librarians will be poised to influence the further development of alternatives to text-based searching. references and notes . bernard j. jansen and amanda spink, “how are we searching the world wide web? a comparison of nine search engine transaction logs,” special issue, information processing and management , no. ( ): . . bernard j. jansen and amanda spink, “an analysis of web documents retrieved and viewed,” in proceedings of the th inter- national conference on internet computing (las vegas, ), . . aravindan veerasamy and nicholas j. belkin, “evaluation of a tool for visualization of information retrieval results,” sigir forum (acm special interest group on information retrieval) ( ): – ; katy börner and javed mostafa, “jodl special issue on information visualization interfaces for retrieval and analysis,” international journal on digital libraries , no. ( ): – ; ozgur turetken and ramesh sharda, “clustering-based visual interfaces for presentation of web search results: an empirical investigation,” information systems frontiers , no. ( ): – . . stephen greene et al., “previews and overviews in digital libraries: designing surrogates to support visual information seeking,” journal of the american society for information science , no. ( ): – ; panayiotis zaphiris et al., “exploring the use of information visualization for digital libraries,” new review of information networking , no. ( ): – . . katy börner and chaomei chen eds., visual interfaces to digital libraries, st ed. (berlin; new york: springer, ), . information technology and libraries | september . zaphiris et al., “exploring the use of information visual- ization for digital libraries,” – . . börner and chen, visual interfaces to digital libraries, . . greene et al., “previews and overviews in digital librar- ies,” – . . “vivisimo corporate profile,” in vivisimo, http://vivi simo.com/html/about (accessed apr. , ). . “aquabrowser library—fiction connection,” www.fic tionconnection.com/ (accessed apr. , ). . “queens library—aquabrowser library,” http://aqua .queenslibrary.org/ (accessed apr. , ). . “xrefer—research mapper,” www.xrefer.com/research (accessed apr. , ). . “stanford ‘groks,’” http://speaking.stanford.edu/back _issues/ soc /library/stanford_groks.html (accessed apr. , ); “grokker at stanford university,” http://library.stan ford.edu/catdb/grokker/ (accessed apr. , ). . “ebsco has partnered with groxis to deliver an inno- vative visual search feature as part of ebsco,” www.groxis .com/service/grokker/pr .html (accessed apr. , ). . michael dolenko, christopher smith, and martha e. williams, “putting the user into usability: developing cus- tomer-driven interfaces at west group,” in proceedings of the national online meeting (medford, n.j.: learned information, ), – ; e. t. morley, “usability testing: the silverplatter experience,” cd-rom professional , no. ( ); ron stew- art, vivek narendra, and axel schmetzke, “accessibility and usability of online library databases,” library hi tech , no. ( ): – ; nicholas tomaiuolo, “deconstructing ques- tia: the usability of a subscription digital library,” searcher , no. ( ): – ; b. hamilton, “comparison of the different electronic versions of the encyclopaedia britannica: a usability study,” electronic library , no. ( ): – ; heather l. munger, “testing the database of international rehabilitation research: using rehabilitation researchers to determine the usability of a bibliographic database,” journal of the medical library association (jmla ) , no. ( ): – ; frank cer- vone, “what we’ve learned from doing usability testing on openurl resolvers and federated search engines,” computers in libraries , no. ( ): – ; alexei oulanov and edmund f. y. pajarillo, “usability evaluation of the city university of new york cuny+ database,” electronic library , no. ( ): – ; steve brantley, annie armstrong, and krystal m. lewis, “usability testing of a customizable library web portal,” college & research libraries , no. ( ): – ; carole a. george, “usability testing and design of a library web site: an iterative approach,” oclc systems & services , no. ( ): – ; leanne m. vandecreek, “usability analysis of northern illinois university libraries’ web site: a case study,” oclc sys- tems & services , no. ( ): – ; susan goodwin, “using screen capture software for web-site usability and redesign buy-in,” library hi tech , no. ( ): – ; laura cobus, valeda frances dent, and anita ondrusek, “how twenty-eight users helped redesign an academic library web site,” refer- ence & user services quarterly , no. ( ): – . . “morae usability testing for software and web sites,” www.techsmith.com/morae.asp (accessed apr. , ). . jansen and spink, “an analysis of web documents retrieved and viewed,” . . ibid. . greene et al., “previews and overviews in digital librar- ies,” . information technology and libraries | june tutorial writing your first scholarly article: a guide for budding authors in librarianship scott nicholson this series of questions and answers is designed to help you take the first steps toward the successful production of a schol- arly article in librarianship. you may find yourself in a library position that requires writing or you may have just decided that you are ready to share your findings, expe- riences, and knowledge with the current and future generations of librarians. while following the guidelines listed here will not guarantee that you will be successful, these steps will take you closer to discovering the thrill of seeing your name in print and making a difference in the field. what should i write about? perhaps you already have an idea based upon your experiences and expertise, or perhaps you aren’t sure which of those ideas you should write about. the best way to start writing is to read other articles! many scholarly articles end with a future research section that outlines other projects and questions that the article sug- gests. it is useful to contact the author of a piece that holds a future research seed to ensure that the author has not already taken on that challenge. sometimes, the original author may be interested in collaborating with you to explore that next question. how do i start? scholarship is an iterative process, in that works that you produce are bricks in an ever-rising wall. your brick will build upon the works of others and, once published, others will build upon your work. because of this, it is essential to begin with a review of related literature. search in bibliographic and citation databases as well as web search tools to see if others have done similar projects to your own. the advantage of finding related literature is that you can learn from the mistakes of others and avoid duplicating works (unless your plan is to replicate the work of others). starting with the work of others allows you to place your brick on the wall. if you do not explicitly discuss how your scholarship relates to the scholarship of others, only those hav- ing familiarity with the literature will be able to understand how your work fits in with that of previous authors. in addition, it’s easier to build upon your work if those who read it have a better idea of the scholarly landscape in which your work lives. as you go out and discover lit- erature, it is crucial to keep citation information about each item. much of what you will cite will be book chap- ters or articles in journals, and you will save yourself time and trouble later if you make a printed copy of source items and record bibliographic information on that copy. recording the title of the work, the full names (including middle initials) of authors and editors, page range, volume, issue, date, publisher and place of publication, url and date accessed, and any other bibliographic infor- mation at the time of collection will save you headaches later when you have to create your references list. as different journals have different cita- tion requirements, having all of this information allows you the flexibility of adapting to different styles. one type of scholarship produced by libraries is the “how our library did something well” article. while a case study of your library can be an appro- priate area of discussion, it is critical to position these pieces within the schol- arship of the field. this allows readers to better understand how applicable your findings are to their own libraries. the concept illustrates the difference between the practice of librarianship and library science. library science is the study of librarianship and includes the generalization of library practice in one setting to other settings. before starting your writing, talk about your idea with your colleagues, which will help you refine your ideas. it will also generate some excitement and publicity about your work, which can help inspire you to continue in the writing process. colleagues can help you consider different places where similar works may already exist and might even open your eyes to similar work in another discipline. you may find a colleague who wants to coauthor the piece with you, which can make the project easier to complete and richer through the collaborative process. another important early step is to consider the journals you would like to be published in. many times, it can be fruitful to publish in the journal that has published works that are in your literature review. considering the journal at this point will allow you to correctly focus the scope, length, and style of your article to the requirements of your desired journal. your article should match the length and tone of other articles in that journal. most journals provide instructions to authors in each issue or on the web; the information page for ital authors is at www.ala.org/ala/ lita/litapublications/ital/information authors.htm. how can i find funding for my research? some projects can’t be easily done in your spare time and require resources for surveys, statistical analysis, travel, or other research costs. you will find that successful requests for funding scott nicholson (srnichol@syr.edu) is an assistant professor in the school of information studies, syracuse university, new york. writing your first scholarly article | nicholson start with a literature review and a research plan. developing these before requesting funding will make your request for funding much stron- ger, as you will be able to demonstrate how your work will sit within a larger context of scholarship. you will need to develop a budget for your funding request. this budget will come together more easily if you have planned out your research. it may be useful or even required for you to develop a set of outcomes for your project and how you will be assessing those outcomes (find more information on outcome-based evaluation through the imls web site at www.imls.gov/grants/current/ crnt_obe.htm). developing this plan will give you a more concrete idea of what resources you will need and when, as well as how you can use the results of your work. resources for research may come from the inside, such as the library or the parent organization of the library, or from an external source, such as a granting body or a corporate donor. in choosing an organization for selec- tion, you should consider who would most benefit from the research, as the request for funding should focus on the benefit to the granting body. many libraries and schools do have small pots of money available for research that will benefit that institution and that, many times, go untapped due to a lack of interest. granting organizations put out formal calls for grant proposals. these can result in a grant that would carry some prestige but would require a detailed formal application that can take months of writing and waiting. another approach is to work with a corporate or nonprofit organization that gives grants. if your organization has a development office, this office may be able to help connect you with a potential supporter of your work. how do i actually do the research? just as the most critical part of a dissertation is the proposal, a good research plan will make your research process run smoothly. before you start the research, write the literature review and the research plan as part of an article. it can be useful to create tables and charts with dummy data that will show how you plan to pres- ent results. doing this allows you to notice gaps in your data-collection plan well before you start that pro- cess. in many research projects, you only have a single chance to collect data; therefore, it’s important to plan out the process before you begin. how do i start writing the paper? the best way to start the writing process is to just write. don’t worry about coming up with a title; the title will develop as the work develops. you can skip over the abstract and introduction; these can be much easier to write after the main body of the article is complete. if you’ve fol- lowed the advice in this paper, then you’ve already written a literature review and perhaps a research plan; these make a good starting point for your article. one way to develop the body of the article is to develop an outline of headings and subheadings. starting with this type of outline forces you to think through your entire article and can help you identify holes in your preparation. once you have the outline completed, you can then fill in the outline by adding text to the head- ings and subheadings. this approach will keep your thinking organized in a way typically used in scholarly writing. scholarly writing is different than creative writing. many librarians with a humanities background face some challenges in transitioning to a differ- ent writing style. scholarly writing is terse; strunk and white’s the elements of style ( ) focuses on succinct writing and can help you refresh your writing skills. if you are having difficulty finding the time to write, it can be useful to set a small quota of writing that you will do every day. a quota such as four paragraphs a day is a reasonable amount to fit into even a busy day, but it will result in the completion of your first draft in only a few weeks. i’m finished with my first complete draft! now what? while you will be excited with the completion of the draft, it’s not appro- priate to send that off to a journal just yet. take a few days off and let your mind settle from the writing, then go back and reread your article carefully. examine each sentence for a subject and a verb, and remove unneeded words, phrases, sentences, para- graphs, or even pages. try to tighten and clean your writing by replacing figures of speech with statements that actually say what you mean in that situation and removing unneeded references to first- and second-per- son pronouns. working through the entire article in this way greatly improves your writing and reduces the review and editing time needed for the article. after this, have several colleagues read your work. some of these might be people with whom you shared your original ideas, and others may be new to the concepts. it can be useful to have members of different depart- ments and with different backgrounds read the piece. ask them if they can read your work by a specific date, as this type of review work is easy to put off when work gets busy. these col- leagues may be people who work in your institution or may be people you have met online. if you know nobody who would be appropriate, consider putting out a request for assistance on a library discussion list focused on your research topic. dealing with the comments from others requires you to set aside your information technology and libraries | june defenses. you did spend a lot of time on this work and it can be easy to slip into a defensive mode. attempt to read their comments from an objective viewpoint. remember—these people are spending their time to help you, and a comment you disagree with at first blush may make more sense if you consider the question “why would someone say this about my work?” putting yourself into the reader’s shoes can aid you in the cre- ation of a piece that speaks to many audiences. what goes on when i submit my work? at this point, your readers have looked at the piece, and you have made corrections on it. now you’re ready to submit your work. follow the directions of the target journal, including length, citation format, and method of submission. if submission is made by e-mail, it would be appro- priate to send a follow-up e-mail a few days after submission to ensure the work was received; it can be very frustrating to realize, after a month of waiting, that the editor never got the work. once you have submitted your work, the editor will briefly review it to ensure it is an appropriate sub- mission for the journal. if it is appro- priate, then the editor will pass the article on to one or more reviewers; if not, you will receive a note fairly quickly letting you know that you should pick another journal. if the reviewing process is “blind,” then you will not know who your reviewers are, but they may know your iden- tity. if the process is “double-blind,” neither reviewer nor author will know the identity of the other. the reviewers will read the article and then submit comments and a recom- mendation to the editor. the editor will collect comments from all of the reviewers and put them together, and send those comments to you. this will always take longer than you would prefer; in reality, it will usually take two to six months, depending upon the journal. after a few months, it would be appropriate for you to contact the edi- tor and ask about the progress on the article and when you should expect comments. do not expect to have your article accepted on the first pass. the common responses are: ■ reject. at this point, you can read the comments provided, make changes, and submit it to another journal. ■ revise and resubmit. the journal is not making a commitment to you, but they are willing to take another look if you are willing to make changes. this is a common response for first submissions. ■ accept with major changes. the journal is interested in publish- ing the article, but it will require reworking. ■ accept with minor changes. you will be presented with a series of small changes. some of these might be required and others might be your choice. ■ accept. the article is through the reviewing process and is on to the next stage. this is an iterative process. you will most likely go through several cycles of this before your article is accepted, and staying dedicated to the process is key to its success. it can be disheartening to have made three rounds of changes only to face another round of small changes. ideally, each set of requested changes should be smaller (and take less time) until you reach the acceptance level. do not submit your work to mul- tiple journals at the same time. if you choose to withdraw your work from one journal and submit it to another, let the editor know that you are doing this (assuming they have not rejected your work). my article has been accepted. when will it come out? once your article is accepted, it will be sent into a copyediting process. the copy editor will contact you with more questions that focus more on writing and citation flaws than on content. after making more cor- rections, you will receive a proof to review (usually with a very tight deadline). this proof will be what comes out in the journal, so check important things like your name, institutions, and contact information carefully. the journal will usually come out several months after you see this final proof. the process from acceptance to publication can take from six months to two years (or more), depending on how much of a publication queue the journal has. the editor should be able to give you an estimate as to when the article will come out after full acceptance. can i put a copy of my article online? it depends upon the copyright agree- ment that you sign. many publishers will allow you to put a copy of your article on a local or institutional web site with an appropriate citation. some allow you to put up a preprint, which would be the version after copyediting but not the final proof version. if the copyright agreement doesn’t say anything about this, then ask the editor of the journal about the policy of authors mounting their own articles on a web site. conclusion writing an article and getting it published is akin to having a child. your child will have a life of its own, and others may notice this new piece of knowledge and build upon it to improve their own library services writing your first scholarly article | nicholson or even make their own works. it is a way to make a difference that goes far beyond the walls of your own library, to extend your professional network, and to engage other scholars in the continued development of the knowledge base of our field. reference . w. strunk jr. and e. b. white, the elements of style (boston: allyn & bacon, ). for more information: w. crawford, first have something to say: writing for the library profession (chicago: ala, ). r. gordon, the librarian’s guide to writing for publication (lanham, md.: scarecrow, ). l. hinchliffe and j. dorner, eds., how to get published in lis journals: a practical guide (san diego: elsevier, ), www .elsevier.com/framework_librarians/lib raryconnect/lcpamphlet .pdf, (accessed feb. , ). editorial the authors of “the state of rfid applications in libraries,” that appeared in the march issue, inadvertently included two sentences that are near quotations from a commentary by peter warfield and lee tien in the april , issue of the berkeley daily planet. on page immediately following footnote , the authors wrote: “the eugene public library reported ‘collision’ problems on very thin materials and on videos as well as false readings from the rfid security gates. collision problems mean that two or more tags are close enough to cancel the signals, making them undetectable by the rfid checkout and security systems.” warfield and lien wrote: “the eugene (ore.) public library reported ‘collision’ problems on very thin materials and on videos as well as ‘false readings’ from the rfid security gates. (collision prob- lems mean that two or more tags are close enough to ‘cancel the signals,’ according to an american library association publication, making them undetectable by the rfid check- out and security systems.)” (accessed may , , www .berkeleydailyplanet.com/article.cfm?archivedate= - - &storyid= ). the authors’ research notes indicated that it was a near quotation, but this fact was lost in the writing of the article. the article referee, the copy editors, and i did not question the authors because earlier in the same paragraph they wrote about the eugene public library experience and referred (footnote ) to an earlier article in the berkeley daily planet. the authors and i apologize for this unfortunate error. **** july , marked the merger of rlg and oclc. by the time this editorial appears, many words will already have been spoken and written about this monumental, twenty- first century library event. i know what i think the three very important immediate effects of the merger will be. first, it is a giant step toward the realization of a global library bibliographic database. second, taking advantage of rlg’s unique and successful programs and integrat- ing them and their development philosophy as “rlg- programs,” while working alongside oclc research, seems a step so important for the future development of library technology that it cannot be overemphasized. third, and very practically, incorporating redlightgreen into open worldcat will give the library world a product that users might prefer over a search of google books or amazon. i requested and received quotes about the merger from the principals that i might put into this editorial that won’t appear until four months after the may announce- ment. jay jordan, president and ceo, oclc, remarked: “we have worked cooperatively with rlg on a variety of projects over the years. since we announced our plans to combine, staff from both organizations have been work- ing together to develop plans and strategies to integrate systems, products, and services. over the past several months, staff members have demonstrated great mutual respect, energy, and enthusiasm for the potential of our new relationship and what it means for the organizations we serve. there is much work to be done as we complete this transition. clearly, we are off to a good start.” betsy wilson, chair, oclc board of trustees, and dean of libraries, university of washington, wrote: “the response from our constituencies has been overwhelmingly supportive. over the past several months, we have final- ized appointments for the twelve-person program council, which reports to . . . oclc through a standing committee called the rlg board committee. we are starting to build agendas for our new alliance. the members of this group from the rlg board are: james neal, vice president for information services and university librarian, columbia university; nancy eaton, dean of university libraries and scholarly communication, penn state university (and for- mer chair of the oclc board); and carol mandel, dean of libraries, new york university. from oclc the members are elisabeth niggeman, director, deutschesbibliothek; jane ryland, senior scientist, internet ; and betsy wilson, dean of university libraries, university of washington.” and from james michalko, currently president and ceo of rlg, and by the time you read this, vice president, rlg-programs development, oclc: “we are combining the practices of rlg and oclc in a very powerful way— by putting together the traditions of rlg and oclc we are creating a robust new venue for research institutions and new capacity that will provide unique and beneficial outcomes to the whole community.” by now, all lita members and ital readers know that in , fred kilgour founded oclc; and was the found- ing editor of the journal of library automation (jola—vol. , no. was published in march, ), which, with but a mild outcry from serials librarians, changed its title to information technology and libraries in . this afternoon ( / / ), i called fred. he and his wife eleanor remi- nisced about the earliest days, and then i asked him for his comments on the oclc-rlg merger. because he had had the first words about both oclc and jola, as it were, i told him that i would like for him to have the last. and this is what he said, “at long last!” fred kilgour died on july , , aged . a tribute posted by alane wilson of oclc may be read at http:// scanblog.blogspot.com/ / /frederick-g-kilgour - - .html editorial: a confession, a speculation, and a farewell john webb john webb (jwebb@wsu.edu) is a librarian emeritus, washington state university and editor of information technology and libraries. editorial | webb bailey information technology and libraries | september three critical issues—a dramatic expansion of the scope, duration, and punitive nature of copyright laws; the abil- ity of digital rights management (drm) systems to lock-down digital content in an unprecedented fashion; and the erosion of net neutrality, which ensures that all internet traffic is treated equally—are examined in detail and their potential impact on libraries is assessed. how legislatures, the courts, and the commercial marketplace treat these issues will strongly influence the future of digital information for good or ill. editor's note: this article was submitted in honor of the fortieth anniversaries of lita and ital. b logs. digital photo and video sharing. podcasts. rip/mix/burn. tagging. vlogs. wikis. these buzz- words point to a fundamental social change fueled by cheap personal computers (pcs) and servers, the internet and its local wired/wireless feeder networks, and powerful, low-cost software. citizens have morphed from passive media consumers to digital-media producers and publishers. libraries and scholars have their own set of buzzwords: digital libraries, digital presses, e-prints, institutional re- positories, and open-access (oa) journals, to name a few. they connote the same kind of change: a democratiza- tion of publishing and media production using digital technology. it appears that we are on the brink of an exciting new era of internet innovation: a kind of digital utopia. gary flake of microsoft has provided one striking vision of what could be (with a commercial twist) in a presentation entitled “how i learned to stop worrying and love the imminent internet singularity,” and there are many other visions of possible future internet advances. when did this metamorphosis begin? it depends on who you ask. let’s say the late s, when the internet began to get serious traction and an early flowering of noncommercial digital publishing occurred. in the subsequent twenty-odd years, publishing and media production went from being highly centralized, capital-intensive analog activities with limited and well- defined distribution channels, to being diffuse, relatively low-cost digital activities with the global internet as their distribution medium. not to say that print and conven- tional media are dead, of course, but it is clear that their era of dominance is waning. the future is digital. nor is it to say that entertainment companies (e.g., film, music, radio, and television companies) and information companies (e.g., book, database, and serial publishers) have ceded the digital-content battlefield to the upstarts. quite the contrary. high-quality, thousand-page-per-volume scientific jour- nals and hollywood blockbusters cannot be produced for pennies, even with digital wizardry. information and enter- tainment companies still have an important role to play, and, even if they didn’t, they hold the copyrights to a significant chunk of our cultural heritage. entertainment and information companies have under- stood for some time that they must adapt to the digital environment or die, but this change has not always been easy, especially when it involves concocting and embracing new business models. nonetheless, they intend to thrive and prosper—and to do whatever it takes to succeed. as they should, since they have an obligation to their share- holders to do so. the thing about the future is that it is rooted in the past. culture, even digital culture, builds on what has gone before. unconstrained access to past works helps determine the richness of future works. inversely, when past works are inaccessible except to a privileged minority, future works are impoverished. this brings us to a second trend that stands in opposi- tion to the first. put simply, it is the view that intellectual works are property; that this property should be protected with the full force of civil and criminal law; that creators have perpetual, transferable property rights; and that contracts, rather than copyright law, should govern the use of intellectual works. a third trend is also at play: the growing use of digital rights management (drm) technologies. when intel- lectual works were in paper (or other tangible forms), they could only be controlled at the object-ownership or object-access levels (a library controlling the circulation of a copy of a book is an example of the second case). physical possession of a work, such as a book, meant that the user had full use of it (i.e., the user could read the entire book and photocopy pages from it). when works are in digital form and are protected by some types of drm, this may no longer be true. for example, a user may only be able to view a single chapter from a drm-protected e-book and may not be able to print it. the fourth and final trend deals with how the internet functions at its most fundamental level. the internet was designed to be content-, application-, and hardware-neu- tral. as long as certain standards were met, the network did not discriminate. one type of content was not given preferential delivery speed over another. one type of strong copyright + drm + weak net neutrality = digital dystopia? charles w. bailey jr. charles w. bailey jr. (cbailey@digital-scholarship.com) is assistant dean for digital library planning and development at university of houston libraries. digital dystopia | bailey content was not charged for delivery while another was free. one type of content was not blocked (at least by the network) while another was unhindered. in recent years, network neutrality has come under attack. the collision of these trends has begun in courts, leg- islatures, and the marketplace. it is far from over. as we shall see, its outcome will determine what the future of digital culture looks like. ฀ stronger copyright: versus copyright law is a complex topic. it is not my intention to provide a full copyright primer here. (indeed, i will assume that the reader understands some copyright basics, such as the notion that facts and ideas are not cov- ered by copyright.) rather, my aim is to highlight some key factors about how and why united states copyright law has evolved and how it relates to the digital problem at hand. three authors (lawrence lessig, professor of law at the stanford law school; jessica litman, professor of law at the wayne state university law school; and siva vaidhyanathan, assistant professor in the department of culture and communication at new york university) have done brilliant and extensive work in this area, and the following synopsis is primarily based on their con- tributions. i heartily recommend that you read the cited works in full. the purpose of copyright let us start with the basis of u.s. copyright law, the constitution’s “progress clause”: “congress has the power to promote the progress of science and useful arts, by securing for limited times to authors and inventors the exclusive right to their respective writings and discoveries.” copyright was a bargain: society would grant creators a time-limited ability to control and profit from their works before they fell into the public domain (where works are unprotected) because doing so resulted in “progress of science and useful arts” (a social good). regarding the progress clause, lessig notes: it does not say congress has the power to grant “creative property rights.” it says that congress has the power to promote progress. the grant of power is its purpose, and its purpose is a public one, not the purpose of enriching publishers, nor even primarily the purpose of reward- ing authors. however, entertainment and information companies can have a far different view, as illustrated by this quote from jack valenti, former president of the motion picture association of america: “creative property owners must be accorded the same rights and protections resident in all other property owners in the nation.” types of works covered when the copyright act of was enacted, it protected published books, maps, and charts written by living u.s. authors as well as unpublished manuscripts by them. the act gave the author the exclusive right to “print, reprint, publish, or vend” these works. now, copyright protects a wide range of published and unpublished “original works of authorship” that are “fixed in a tangible medium of expression” without regard for “the nationality or domi- cile of the author,” including “ . literary works; . musical works, including any accompanying words; . dramatic works, including any accompanying music; . pantomimes and choreographic works; . pictorial, graphic, and sculp- tural works; . motion pictures and other audiovisual works; . sound recordings; . architectural works.” rights in contrast to the limited print publishing rights inherent in the copyright act of , current law grants copyright owners the following rights (especially notable is the addi- tion of control over derivative works, such as a play based on a novel or a translation): ฀ to reproduce the work in copies or phonograph records; ฀ to prepare derivative works based upon the work; ฀ to distribute copies or phonograph records of the work to the public by sale or other transfer of ownership, or by rental, lease, or lending; ฀ to perform the work publicly, in the case of literary, musi- cal, dramatic, and choreographic works, pantomimes, and motion pictures and other audiovisual works; ฀ to display the copyrighted work publicly, in the case of literary, musical, dramatic, and choreographic works, pantomimes, and pictorial, graphic, or sculptural works, including the individual images of a motion picture or other audiovisual work; and ฀ in the case of sound recordings, to perform the work pub- licly by means of a digital audio transmission. duration the copyright act of granted authors a term of four- teen years, with one renewal if the author was still living (twenty-eight years total). now the situation is much more complex, and, rather than trying to review the details, i’ll provide the following example. for a personal author who produced a work on or after january , , it is covered for the life of the author plus seventy years. so, assuming information technology and libraries | september an author lives an average seventy-five years, the work would be covered for years, which is approximately years longer than in . registration registration was required by the copyright act of , but very few eligible works were registered from to , which enriched the public domain. now registra- tion is not required, and no work enriches the public domain until its term is over, even if the author (or the author’s descendants) have no interest in the work being under copyright, or it is impossible to locate the copyright holder to gain permission to use his or her works (creating so-called “orphan works”). drafting of legislation by , copyright law had become fairly esoteric and com- plex, and drafting new copyright legislation had become increasingly difficult. consequently, congress adopted a new strategy: let those whose commercial interests were directly affected by copyright law deliberate and negoti- ate with each other about copyright law changes, and use the results of this process as the basis of new legislation. over time, this increasingly became a dialogue among representatives of entertainment, high-tech, information, and telecommunications companies; other parties, such as library associations; and rights-holder groups (e.g., ascap). since these parties often had competing interests, the negotiations were frequently contentious and lengthy. the resulting laws created a kind of crazy quilt of specific exceptions for the deals made during these sessions to the ever-expanding control over intellectual works that copyright reform generally engendered. since the public was not at the table, its highly diverse interests were not directly represented, and, since stakeholder industries lobby congress and the public does not, the public’s interests were often not well served. (there were some efforts by special interest groups to represent the public on narrowly focused issues.) frequency of copyright term legislation with remarkable restraint, congress, in its first hundred years, enacted one copyright bill that extended the copy- right term and one in its next fifty; however, starting in , it passed eleven bills in the next forty years. famously, jack valenti once proposed that copyright “last forever less one day.” by continually extending copyright terms in a serial fashion, congress may grant him his wish. licenses in , copyrighted works were sold and owned. today, many digital works are licensed. licenses usually fall under state contract law rather than federal copyright law. licensed works are not owned, and the first-sale doctrine is not in effect. while copyright is the legal foundation of licenses (i.e., works can be licensed because licensors own the copyright to those works), licenses are contracts, and contract provisions trump user-favorable copyright provisions, such as fair use, if the licensor chooses to negate them in a license. criminal and civil penalties in there were civil penalties for copyright infringe- ment (e.g., statutory fines of “ cents per sheet found in the infringer ’s possession”). now there are criminal copyright penalties, including felony violations that can result in a maximum of five years of imprisonment and fines as high as $ , for first-time offenders; civil statutory fines that can range as high as $ , per infringement (if infringement is “willful”), and other penalties. once the copyright implications of digital media and the internet sunk in, entertainment and information companies were deeply concerned: digital technologies made creating perfect copies effortless, and the internet provided a free (or low-cost) way to distribute content globally. congress, primarily spurred on by entertainment companies, passed several laws aimed at curtailing perceived digital “theft” through criminal penalties. under the no electronic theft (net) act, copyright infringers face “up to years in prison and/or $ , fines,” even for noncommercial infringement. under the digital millennium copyright act (dmca), those who defeat technological mechanisms that control access to copyrighted works (a process called “circum- vention”) face a maximum of five years in prison and $ , in fines. effect of copyright on average citizens in , copyright law had little effect on citizens. the average person was not an author or publisher, private use of copyrighted materials was basically unregulated, the public domain was healthy, and many types of works were not covered by copyright at all. in , ฀ virtually every type of work imaginable is under automatic copyright protection for extended periods of time; ฀ private use of digital works is increasingly visible and of concern to copyright holders; ฀ the public domain is endangered; and ฀ ordinary citizens are being prosecuted as “pirates” under draconian statutory and criminal penalties. digital dystopia | bailey regarding this development, lessig says: for the first time in our tradition, the ordinary ways in which individuals create and share culture fall within the reach of the regulation of the law, which has expanded to draw within its control a vast amount of culture and creativity that it never reached before. the technology that preserved the balance of our history—between uses of our culture that were free and uses of our culture that were only upon permission—has been undone. the consequence is that we are less and less a free culture, more and more a permission culture. how has copyright changed since the days of the founding fathers? as we have seen, there has been a shift in copyright law (and social perceptions of it) from ฀ promoting progress to protecting intellectual prop- erty owners’ “rights”; ฀ from covering limited types of works to covering virtually all types of works; ฀ from granting only basic reproduction and distribu- tion rights to granting a much wider range of rights; ฀ from offering a relatively short duration of protection to offering a relatively long (potentially perpetual) one; ฀ from requiring registration to providing automatic copyright; ฀ from drafting laws in congress to drafting laws in work groups of interested parties dominated by com- mercial representatives; ฀ from making infrequent extensions of copyright duration to making frequent ones; ฀ from selling works to licensing them; ฀ from relatively modest civil penalties to severe civil and criminal penalties; and ฀ from ignoring ordinary citizens’ typical use of copy- righted works to branding them as pirates and prosecut- ing them with lawsuits. (regarding lawsuits filed by the recording industry association of america against four students, lessig notes: “if you added up the claims, these four lawsuits were asking courts in the united states to award the plaintiffs close to $ billion—six times the total profit of the film industry in .”) complicating this situation further is intense consolida- tion and increased vertical integration in the entertainment, information, telecommunications, and other high-tech industries involved in the internet. this vertical integra- tion has implications for what can be published and the free flow of information. for example, a company that publishes books and magazines, produces films and television pro- grams, provides internet access and digital content, and provides cable television services (including broadband internet access) has different corporate interests than a company that performs a single function. these interrelated interests may affect not only what information is produced and whether competing information and services are freely available through controlled digital distribution channels, but corporate perceptions of copyright issues as well. one of the ironies of the current copyright situation is this: if creative works are by nature property, and stealing property is (and has always been) wrong, then some of the very industries that are demanding that this truth be embodied in copyright law have, in the past, been pirates themselves, even though certain acts of piracy may have been legal (or appeared to be legal) under then-existing copyright laws. lessig states: if “piracy” means using the creative property of others without their permission—if “if value, then right” is true—then the history of the content industry is a his- tory of piracy. every important sector of “big media” today—film, records, radio, and cable tv—was born of a kind of piracy so defined. the consistent story is how last generation’s pirates join this generation’s country club—until now. let’s take a simple case: cable television. early cable television companies used broadcast television programs without compensating copyright owners, who branded their actions as piracy and filed lawsuits. after two defeats in the supreme court, broadcast television companies won a victory (of sorts) in congress, which took nearly thirty years to resolve the matter: cable television companies would pay, but not what broadcast television companies wanted; rather they would pay fees determined by law. of course, this view of history (big media companies as pirates in their infancy) is open to dispute. for the moment, let’s assume that it is true. put more gently, some of the most important media companies of modern times flourished because of relatively lax copyright control, a relatively rich public domain, and, in some cases, a societal boon that allowed them to pay statutory license fees— which are compulsory for copyright owners—instead of potentially paying much higher fees set by copyright owners or being denied use at all. today, the very things that fostered media companies’ growth are under attack by them. the success of those attacks is diminishing the ability of new digital content and service companies to flourish and, in the long run, may diminish even big media’s ability to continue to thrive as a permission culture replaces a permissive culture. several prominent copyright scholars have suggested copyright reforms to help restore balance to the copyright system. james boyle, professor of law at the duke university law school, recommends a twenty-year copyright term with “a broadly defined fair use protection for journalis- tic, teaching, and parodic uses—provided that those uses were not judged to be in bad faith by a jury applying the ‘beyond a reasonable doubt’ standard.” information technology and libraries | september william w. fisher iii, hale and dorr professor of intellectual property law at harvard university law school, suggests that “we replace major portions of the copyright and encryption-reinforcement models with . . . a governmentally administered reward system” that would put in place new taxes and compensate registered copyright owners of music or films with “a share of the tax revenues proportional to the relative popularity of his or her creation,” and would “eliminate most of the current prohibitions on unauthorized reproduction, distribution, adaptation, and performance of audio and video recordings.” lessig recommends that copyright law be guided by the following general principles: ( ) short copyright terms, ( ) a simple binary system of protected/not pro- tected works without complex exceptions, ( ) mandatory renewal, and ( ) a “prospective” orientation that forbids retrospective term extensions. (previously, lessig had proposed a seventy-five-year term contingent on five-year renewals). he suggests reinstating the copyright registra- tion requirement using a flexible system similar to that used for domain name registrations. he favors works having copyright marks, and, if they are not present, he would permit their free use until copyright owners voice their opposition to this use (uses of the work made prior to this point would still be permitted). litman wants a copyright law “that is short, simple, and fair,” in which we “stop defining copyright in terms of reproduction” and recast copyright as “an exclusive right of commercial exploitation.” litman would eliminate indus- try-specific copyright law exceptions, but grant the public “a right to engage in copying or other uses incidental to a licensed or legally privileged use”; the “right to cite” (even infringing works); and “an affirmative right to gain access to, extract, use, and reuse the ideas, facts, information, and other public-domain material embodied in protected works” (including a restricted circumvention right). things change in two hundred-plus years, and the law must change with them. since the late nineteenth century, copyright law has been especially impacted by new tech- nologies. the question is this: has copyright law struck the right balance between encouraging progress through granting creators specific rights and fostering a strong public domain that also nourishes creative endeavor? if that balance has been lost, how can it be restored? or is society simply no longer striving to maintain that balance because intellectual works are indeed property, property must be protected for commerce to prosper, and the concept of bal- ance is outmoded and no longer reflects societal values? ฀ drm: locked-up content and fine-grained control noted attorney michael godwin defines drm as “a collec- tive name for technologies that prevent you from using a copyrighted digital work beyond the degree to which the copyright owner (or a publisher who may not actually hold a copyright) wishes to allow you to use it.” like copyright, drm systems are complex, with many variations. there are two key technologies: ( ) digital mark- ing (i.e., digital fingerprints that uniquely identify a work based on its characteristics, simple labels that attach rights information to content, and watermarks that typically hide information that can be used to identify a work), and ( ) encryption (i.e., scrambled digital content that requires a digital key to decipher it). specialized hardware can be used to restrict access as well, often in conjunction with digital marking and encryption. the intent of this article is not to provide a technical tutorial, but to set forth an overview of the basic drm concept and discuss its implications. what is of interest here is not how system a-b-c works in contrast to system x-y-z, but what drm allows copyright owners to do and the issues related to drm. to do so, let’s use an analogy, understanding that real drm systems can work in other ways as well (e.g., digi- tal watermarks can be used to track illegal use of images on the internet without those images being otherwise protected). for the moment, let’s imagine that the content a user wishes to access is in an unbreakable, encrypted digital safe. the user cannot see inside the safe. by entering the correct digital combination, certain content becomes visible (or audible or both) in the safe. that content can then be utilized in specific ways (and only those ways), including, if permitted, leaving the safe. if a public domain work is put in the safe, access to it is restricted regardless of its copyright status. bill rosenblatt, bill trippe, and stephen mooney pro- vide a very useful conceptual model of drm rights in their landmark drm book, digital rights management: business and technology, summarized here. there are three types of content rights: ( ) render rights, ( ) transport rights, and ( ) derivative-works rights. render rights allow authorized users to view, play, and print protected content. transport rights allow authorized users to copy, move, and loan content (the user retains the content if it is copied and gets it back when a loan is over, but does not keep a copy if it is moved). derivative-works rights allow authorized users to extract pieces of content, edit the content in place, and embed content by extracting some of it and using it in other works. each one of these individual rights has three attributes: ( ) consideration, ( ) extents, and ( ) types of users. in the first attribute, consideration, access to content is provided for something of value to the publisher (e.g., money or personal information). content can then be used to some extent (e.g., for a certain amount of time or a certain number of times). the rights and attributes users have are determined by their user types. digital dystopia | bailey for example, an academic user, in consideration of a specified license payment by his or her library, can view a drm-protected scholarly article—but not copy, move, loan, extract, edit, or embed it—for a week, after which it is inaccessible. we can extend this hypothetical example by imagining that the library could pay higher license fees to gain more rights to the journal in question, and the library (or the user) could dynamically purchase additional article-specific rights enhancements as needed through micropayments. this example is extreme; however, it illustrates the fine-grained, high level of control that publishers could potentially have over content by using drm technology. godwin suggests that drm may inhibit a variety of legitimate uses of drm-protected information, such as access to public-domain works (or other works that would allow liberal use), preservation of works by librar- ies, creation of new derivative works, conduct of histori- cal research, exercise of fair-use rights, and instructional use. the ability of blind (or otherwise disabled) users to employ assistive technologies may also be prevented by drm technology. drm also raises a variety of privacy concerns. fair use is an especially thorny problem. rosenblatt, trippe, and mooney state: fair use is an “i’ll know it when i see it” proposition, meaning that it can’t be proscriptively defined. . . . just as there is no such thing as a “black box” that determines whether broadcast material is or isn’t indecent, there is no such thing as a “black box” that can determine whether a given use of content qualifies as fair use or not. anything that can’t be proscriptively defined can’t be represented in a computer system. no need to panic about scholarly journals—yet. your scholarly journal publisher or other third-party supplier is unlikely to present you with such detailed options tomor- row. but you may already be licensing other digital content that is drm-protected, such as digital music or e-books that require a hardware e-book reader. as the recent sony bmg “rootkit” episode illustrated, creating effective, secure drm systems can be challeng- ing, even for large corporations. again, the reasons for this are complex. in very simple terms, it boils down to this: assuming that the content can be protected up to the point it is placed in a drm system, the drm system has the best chance of working if all possible devices that can process its protected content either directly support its protection technology, recognize its restrictions and enforce them through another means, or refuse access. anything less creates “holes” in the protective drm shell, such as the well-known “analog hole” (e.g., when drm-protected digital content is converted to analog form to be played, it can then be rerecorded using digital equipment without drm protection). ideally, in other words, every server, network router, pc and pc component, operating system, and relevant electronic device (e.g., cd player, dvd player, audio- recording device, and video-recording device) would work with the drm system as outlined previously or would not allow access to the content at all. clearly, this ideal end-state for drm may well never be realized, especially given the troublesome backward- compatibility equipment problem. however, this does not mean that the entertainment, information, and high- technology companies will not try to make whatever piecemeal progress that they can in this area. the trusted computing group is an important mul- tiple-industry security organization, whose standards work could have a strong impact on the future of drm. robert a. gehring notes: but a drm system is almost useless, that is from a con- tent owner’s perspective, until it is deployed broadly. putting together cheap tc components with a market- dominating operating system “enriched” with drm functionality is the most economic way to provide the majority of users with “copyright boxes.” seth schoen argues computer owners should be empowered to override certain features of “trusted com- puting architecture” to address issues with “anti-competi- tive and anti-consumer behavior” and other problems. drm could potentially be legislatively mandated. there is a closely related legal precedent, the audio home recording act, which requires that digital audiotape equip- ment include special hardware to prevent serial copying. there is currently a bill before congress that would require use of a “broadcast flag” (a digital marker) for digital broadcast and satellite radio receivers. last year, a similar fcc regulation for broadcast digital television was struck down by a federal appeals court; consequently, the current bill explicitly empowers the fcc to “enforce ‘prohibi- tions against unauthorized copying and redistribution.’” another bill would plug the analog-to-digital video analog hole by putting “strict legal controls on any video analog to digital (a/d) convertors.” whether these bills become law or not, efforts to mandate drm are unlikely to end. dmca strongly supports drm by prohibiting both the circumvention of technological mechanisms that control access to copyrighted works (with some minor exceptions) and the “manufacture of any device, composition of any program, or offering of any service” to do so. what would the world be like if all newly published (or released) commercially created information was in digital form, protected by drm? what would it be like if all old works in print and analog formats were only reissued in digital form, protected by drm? what would it be like if all hardware that could process that digital information had to support the information’s drm scheme or block any access to it because this was mandated by law? what would it be information technology and libraries | september like if all operating systems had direct or indirect built-in support for drm? would “progress of science and useful arts” be promoted or squashed? ฀ weaker net neutrality lessig identifies three important characteristics of the internet that have fostered innovation: ( ) edge architec- ture: software applications run on servers connected to the network, rather than on the network itself, ensuring that the network itself does not have to be modified for new or updated applications to run; ( ) no application optimization: a relatively simple, but effective, protocol is utilized (internet protocol) that is indifferent to what software applications run on top of it, again insulating the network from application changes; and ( ) neutral platform: the network does not prefer certain data packets or deny certain packets access. lessig’s conceptual model is very useful when thinking about net neutrality, a topic of growing concern. educause’s definition of net neutrality aptly cap- tures these concerns: “net neutrality” is the term used to describe the concept of keeping the internet open to all lawful content, infor- mation, applications, and equipment. there is increasing concern that the owners of the local broadband connec- tions (usually either the cable or telephone company) may block or discriminate against certain internet users or applications in order to give an advantage to their own services. while the owners of the local network have a legitimate right to manage traffic on their network to pre- vent congestion, viruses, and so forth, network owners should not be able to block or degrade traffic based on the identity of the user or the type of application solely to favor their interests. for some time, there have been fears that net neutral- ity was endangered as the internet became increasingly commercialized, a greater percentage of home internet users migrated to broadband connections not regulated by common carrier laws, and telecommunications mergers (and vertical integration) accelerated. some of these fears are now appearing to be realized, albeit with resistance by the internet community. for example, aol has indicated that it will implement a two-tier e-mail system for companies, nonprofits, and others who send mass mailings: those who pay bypass spam filters, those who don’t pay don’t bypass spam filters. critics fear that free e-mail services will deterio- rate under a two-tier system. facing fierce criticism from the dearaol.com coalition and many others, aol has relented somewhat on the nonprofit issue by offering special treatment for “qualified” nonprofits. a second example is that an analysis of verizon’s fcc filings reveals that “more than % of verizon’s current capacity is earmarked for carrying its service, while all other traffic jostles in the remainder.” content-oriented net companies are worried: leading net companies say that verizon’s actions could keep some rivals off the road. as consumers try to search google, buy books on amazon.com, or watch videos on yahoo!, they’ll all be trying to squeeze into the leftover lanes on verizon’s network. . . . “the bells have designed a broadband system that squeezes out the public internet in favor of services or content they want to provide,” says paul misener, vice-president for global policy at amazon.com. a third example is a comment by william l. smith, bellsouth ‘s chief technology officer, who “told reporters and analysts that an internet service provider such as his firm should be able, for example, to charge yahoo inc. for the opportunity to have its search site load faster than that of google inc.,” but qualified this assertion by indicat- ing that “a pay-for-performance marketplace should be allowed to develop on top of a baseline service level that all content providers would enjoy.” about four months later, at&t announced that it would acquire bellsouth, after which it “will be the local carrier in states covering more than half of the american population.” finally, in a white paper for public knowledge, john windhausen jr. states: this concern is not just theoretical—broadband network providers are taking advantage of their unregulated status. cable operators have barred consumers from using their cable modems for virtual private networks and home networking and blocked streaming video applications. telephone and wireless companies have blocked internet telephone (voip—voice over the internet protocol) traffic outright in order to protect their own telephone service revenues. these and similar examples are harbingers of troubled days ahead for net neutrality. the canary in the net neu- trality mine isn’t dead yet, but it’s getting very nervous. the bottom line? noted oa advocate peter suber analyzes the situation as follows: but now cable and telecom companies want to discrimi- nate, charge premium prices for premium service, and give second-rate service to everyone else. if we relax the principle of net neutrality, then isps could, if they wanted, limit the software and hardware you could con- nect to the net. they could charge you more if you send or receive more than a set number of emails. they could block emails containing certain keywords or emails from people or organizations they disliked, and block traffic to or from competitor web sites. they could make filtered service the default and force users to pay extra for the digital dystopia | bailey wide open internet. if you tried to shop at a store that hasn’t paid them a kickback, they could steer you to a store that has. . . . if companies like at&t and verizon have their way, there will be two tiers of internet service: fast and expensive and slow and cheap (or cheaper). we unwealthy users—students, scholars, universities, and small publishers—wouldn't be forced offline, just forced into the slow lane. because the fast lane would reserve a chunk of bandwidth for the wealthy, the peons would crowd together in what remained, reducing service below current levels. new services starting in the slow lane wouldn't have a fighting chance against entrenched players in the fast lane. think about ebay in , google in , or skype in without the level playing field provided by network neutrality. or think about any oa journal or repository today. is net neutrality a quaint anachronism of the internet’s distant academic/research roots that we would be better off without? would new internet companies and noncom- mercial services prosper better if it was gone, spurring on new waves of innovation? would telecommunications companies (who may be part of larger conglomerates), free to charge for tiered-services, offer us exciting new service offerings and better, more reliable service? ฀ defending the internet revolution sixties icon bob dylan’s line in “the times they are a- changin’”—“then you better start swimmin’ or you’ll sink like a stone”—couldn’t be more apt for those concerned with the issues outlined in this paper. here’s a brief over- view of some of the strategies being used to defend the freewheeling internet revolution. . darknet: j. d. lasica says: “for the most part, the darknet is simply the underground internet. but there are many darknets: the millions of users trading files in the shady regions of usenet and internet relay chat; students who send songs and tv shows to each other using instant messaging services from aol, yahoo, and microsoft; city streets and college campuses where people copy, burn, and share physical media like cds; and the new breed of encrypted dark networks like freenet. . .” we may think of the darknet as simply fostering illegal file swapping by ordinary citizens, but the darknet strategy can also be used to escape government internet censorship, as is the case with freenet use in china. . legislative and legal action: there have been attempts to pass laws to amend or reverse copyright and other laws resulting from the counter-internet-revolution, which have been met by swift, powerful, and generally effective opposition from entertainment companies and other parties affected by these proposed measures. the moral of this story is that these large corporations can afford to pay lobbyists, make campaign contributions, and otherwise exert significant influence over lawmakers, while, by and large, advocates for the other side do not have the same clout. the battle in the courts has been more of a mixed bag; however, there have been some notable defeats for reform advocates, especially in the copyright arena (e.g., eldred v. ashcroft), where most of the action has been. . market forces: when commercial choices can be made, users can vote with their pocketbooks about some internet changes. but, if monopoly forces are in play, such as having a single option for broadband access, the only other choice may be no service. however, as the oa move- ment (described later) has demonstrated, a concerted effort by highly motivated individuals and nonprofit organiza- tions can establish viable new alternatives to commercial services that can change the rules of the game in some cases. companies can also explore radical new business models that may appear paradoxical to pre-internet-era thinking, but make perfect sense in the new digital real- ity. in the long run, the winners of the digital-content wars may be those who are not afraid of going down the internet rabbit hole. . creative commons: copyright is a two-edged sword: it can be used as the legal basis of licenses (and drm) to restrict and control digital information, or it can be used as the legal basis of licenses to permit liberal use of digital information. by using one of the six major creative common licenses (ccl), authors can retain copyright, but significantly enrich society’s collective cultural repository with works that can be freely shared for noncommercial purposes, used, in some cases, for commercial purposes, and used to easily build new derivative creative works. for example, the creative commons attribution license requires that a work is attributed to the author; however, a work can be used for any commercial or noncommercial purpose without permission, including creating derivative works. there are a variety of other licenses, such as the gnu free documentation license, that can be used for similar purposes. . oa: scholars create certain types of information, such as journal articles, without expecting to be paid to do so, and it is in their best interests for these works to be widely read, especially by specialists in their fields. by putting e-prints (electronic preprints or post-prints) of articles on personal home pages or in various types of digital archives (e.g., institutional repositories) in full compliance with copyright law and, if needed, in compli- ance with publisher policies, scholars can provide free global access to these works with minimal effort and at no (or little) cost to themselves. further, a new generation of free e-journals are being published on the internet that are being funded by a variety of business models, such as advertising, author fees, library membership fees, and supplemental products. these oa strategies make digital information technology and libraries | september scholarly information freely available to users across the globe, regardless of their personal affluence or the affluence of their affiliated institutions. ฀ impact on libraries this paper’s analysis of copyright, drm, and network neutrality trends holds no good news for libraries. copyright the reach of copyright law constantly encompasses new types of materials and for an ever-lengthening duration. as a result, copyright holders must explicitly place their works in the public domain if the public domain is to continue to grow. needless to say, the public domain is a primary source of materials that can be digitized without having to face a complex, potentially expensive, and sometimes hope- less permission clearance process. this process can be especially daunting for media works (such as films and video), even for the use of very short segments of these works. j. d. lasica recounts his effort to get permission to use short music and film segments in a personal video: five out of seven music companies declined; six out of seven movie studios declined, and the one that agreed had serious reservations. the replies to his inquiry, for those companies that bothered to reply at all, are well worth reading. for u.s. libraries without the resources to deal with complicated copyright-related issues, the digitization clock stops at , the last year we can be sure that a work is in the public domain without checking its copyright status and getting permission if it is under copyright. what can we look forward to? lessig says: “thus, in the twenty years after the sonny bono act, while one million patents will pass into the public domain, zero copyrights will pass into the public domain by virtue of the expiration of a copyright term.” (the sonny bono term extension act was passed in .) digital preservation is another area of concern in a legal environment where most information is automati- cally copyrighted, copyright terms are lengthy (or end- less), and information is increasingly licensed. simply put, a library cannot digitally preserve what it does not own unless the work is in the public domain, the work’s license permits it, or the work’s copyright owner grants permission to do so. or can it? after all, the internet archive does not ask permission ahead of time before preserving the entire internet, although it responds to requests to restrict infor- mation. and that is why the internet archive is currently being sued by healthcare advocates, which says that it: “is just like a big vacuum cleaner, sucking up information and making it available.” if it is not settled out of court, this will be an interesting case for more digitally adventurous libraries to watch. as the cost of the hardware and software needed to effectively do so continues to drop, faculty, students, and other library users will increasingly want to repurpose content, digitizing conventional print and media materials, remixing digital ones, and/or creating new digital materi- als from both. with the “information commons” movement, academic libraries are increasingly providing users with the hard- ware and software tools to repurpose content. given that the wording of the u.s. copyright act section (f) ( ) is vague enough that it could be interpreted to include these tools when they are used for information reproduction, is the old “copyright disclaimer on the photocopier” solution enough in the new digital environment? or—in light of the unprecedented transformational power of these tools to create new digital works, and their widespread use both within libraries and on campus—do academic libraries bear heavier responsibilities regarding copyright compli- ance, permission-seeking, and education? similar issues arise when faculty want to place self-cre- ated digital works that incorporate copyrighted materials in electronic reserves systems or institutional repositories. end- user contributions to “library . ” systems that incorporate copyrighted materials may also raise copyright concerns. drm as libraries realize that they cannot afford dual formats, their new journal and index holdings are increasingly solely digital. libraries are also licensing a growing variety of “born digital” information. the complexities of dealing with license restrictions for these commercial digital prod- ucts are well understood, but imagine if drm was layered on top of license restrictions. as we have discussed, drm will allow content producers and distributors to slice, dice, and monetize access to digital information in ways that were previously impossible. what may be every publisher/vendor’s dream could be every library’s nightmare. aside from a potential surge of publisher/vendor-specific access licensing options and fees, libraries may also have to contend with publisher/ vendor-specific drm technical solutions, which may: ฀ depend on particular hardware/software platforms, ฀ be incompatible with each other, ฀ decrease computer reliability and security, ฀ eliminate fair or otherwise legal use of drm-pro- tected information, ฀ raise user privacy issues, ฀ restrict digital preservation to bitstream preservation (if allowed by license), digital dystopia | bailey ฀ make it difficult to assess whether to license drm- protected materials, ฀ increase the difficulty of providing unified access to information from different publishers and vendors, ฀ multiply user support headaches, and ฀ necessitate increased staffing. drm makes solving many of these problems both legally and technically impossible. for example, under dmca, libraries have the right to circumvent drm for a work in order to evaluate whether they want to purchase it. however, they cannot do so without the software tools to crack the work’s drm protection. but the distribution of those tools is illegal under dmca, and local develop- ment of such tools is likely to be prohibitively complex and expensive. fostering alternatives to restrictive copyright and drm given the uphill battle in the courts and legislatures, ccls (or similar licenses) and oa are particularly prom- ising strategies to deal with copyright and drm issues. copyright laws do not need to change for these strategies to be effective. it is not just a question of libraries helping to support oa by paying for institutional memberships to oa jour- nals, building and maintaining institutional repositories, supporting oa mandates, encouraging faculty to edit and publish oa journals, educating faculty about copyright and oa issues, and encouraging them to utilize ccls (or similar licenses). to truly create change, libraries need to “walk the talk” and either let the public-domain materials they digitize remain in the public domain, or put them under ccls (or similar licenses), and, when they create original digital content, put it under ccls (or similar) licenses as well. as the oa movement has shown, using ccls does not rule out revenue generation (if that is an appropriate goal), but it does require facilitating strategies, such as advertis- ing and offering fee-based add-on products and services. net neutrality there are many unknowns surrounding the issue of net neutrality, but what is clear is that it is under assault. it is also clear that internet services are more likely to require more, not less, bandwidth in the future as digital media and other high-bandwidth applications become more com- monplace, complex, and interwoven into a larger number of internet systems. one would imagine that if a corporation such as google had to pay for a high-speed digital lane, it would want it to reach as many consumers as possible. so, it may well be that libraries’ google access would be unaffected or possibly improved by a two-tier (or multi-tier) internet “speed-lane” service model. would the same be true for library-oriented publishers and vendors? that may depend on their size and relative affluence. if so, the ability of smaller publishers and vendors to offer innovative bandwidth-intensive products and services may be curtailed. unless they are affluent, libraries may also find that they are confined to slower internet speed lanes when they act as information providers. for libraries engaged in digital library, electronic publishing, and institutional repository projects, this may be problematic, especially as they increasingly add more digital media, large-data-set, or other bandwidth-intensive applications. it’s important to keep in mind that net neutrality impacts are tied to where the chokepoints are, with the most serious potential impacts being at chokepoints that affect large numbers of users, such as local isps that are part of large corporations, national/international back- bone networks, and major internet information services (e.g.,yahoo!). it is also important to realize that the problem may be partitioned to particular network segments. for example, on-campus network users may not experience any speed issues associated with the delivery of bandwidth-intensive information from local library servers because that net- work segment is under university control. remote users, however, including affiliated home users, may experience throttled-down performance beyond what would normally be expected due to speed-lane enforcement by backbone providers or local isps controlled by large corporations. likewise, users at two universities connected by a spe- cial research network may experience no issues related to accessing the other university’s bandwidth-intensive library applications from on-campus computers because the backbone provider is under a contractual obligation to deliver specific network performance levels. although the example of speed lanes has been used in this examination of potential net neutrality impacts on libraries, the problem is more complex than this, because network services, such as peer-to-peer networking proto- cols, can be completely blocked, digital information can be blocked or filtered, and other types of fine-grained network control can be exerted. ฀ conclusion this paper has deliberately presented one side of the story. it should not be construed as saying that copyright law should be abolished or violated, that drm can serve no useful purpose (if it is possible to fix certain critical defi- ciencies and if it is properly employed), or that no one has to foot the bill for content creation/marketing/distribution and ever-more-bandwidth-hungry internet applications. information technology and libraries | september nor is it to say that the other side of the story, the side most likely to be told by spokespersons of the entertain- ment, information, and telecommunications industries, has no validity and does not deserve to be heard. however, that side of the story is having no problem being heard, especially in the halls of congress. the side of the story presented in this paper is not as widely heard—at least, not yet. nor does it intend to imply that executives from the entertainment, information, telecommunications, and other corporate venues lack a social conscience, are fully unified in their views, or are unconcerned with the societal implications of their positions. however, by focusing on short-term issues, they may not fully realize the potentially negative, long-term impact that their positions may have on their own enterprises. nor has this paper presented all of the issues that threaten the internet, such as assaults on privacy, increas- ingly determined (and malicious) hacking, state and other censorship, and the seemingly insolvable problem of over- laying national laws on a global digital medium. what this paper has said is simply this: three issues—a dramatic expansion of the scope, duration, and punitive nature of copyright laws; the ability of drm to lock-down content in an unprecedented fashion; and the erosion of net neutrality—bear careful scrutiny by those who believe that the internet has fostered (and will continue to foster) a digital revolution that has resulted in an extraordinary explosion of innovation, creativity, and information dis- semination. these issues may well determine whether the much-touted information superhighway lives up to its promise or simply becomes the “information toll road” of the future, ironically resembling the pre-internet online services of the past. references and notes . gary flake, “how i learned to stop worrying and love the imminent internet singularity,” http://castingwords.com/ transcripts/o / .html (accessed may , ). . lawrence lessig, free culture: the nature and future of creativity (new york: penguin, ), , www.free-culture.cc/ (accessed may , ). . ibid., . . ibid., – . . william f. patry, copyright law and practice (washing- ton, d.c.: bureau of national affairs, ), http://digital-law -online.info/patry (accessed may , ). . u.s. copyright office, copyright basics (washington, d.c.: u.s. copyright office, ), www.copyright.gov/circs/circl/ html (accessed may , ). . ibid. . lessig, free culture, . . barbara m. waxer and marsha baum, internet surf and turf revealed: the essential guide to copyright, fair use, and find- ing media (boston: thompson course technology, ), . . patry, copyright law and practice; lessig, free culture, . . jessica litman, digital copyright (amherst: prometheus books, ), – . . lessig, free culture, . . ibid., . . association of american universities, the association of research libraries, the association of american university presses, and the association of american publishers, campus copy- right rights & responsibilities: a basic guide to policy considerations (association of american universities, the association of research libraries, the association of american university presses, and the association of american publishers, ), , www.arl.org/info/ frn/copy/campuscopyright .pdf (accessed may , ). . george h. pike, “the delicate dance of database li- censes, copyright, and fair use,” computers in libraries , no. ( ): , http://infotoday.com/cilmag/may /pike .htm (accessed may , ). . patry, copyright law and practice. . computer crime and intellectual property section crimi- nal division, u.s. department of justice, “prosecuting intellec- tual property crimes manual,” www.cybercrime.gov/ipmanual .htm (accessed may , ); u.s. copyright office, copyright law of the united states of america and related laws contained in title of the united states code (washington, d.c.: u.s. copyright office, ), www.copyright.gov/title /circ .pdf (accessed may , ). . recording industry association of america, “copyright laws,” www.riaa.com/issues/copyright/laws.asp (accessed may , ). . kenneth d. crews, copyright law for librarians and educa- tors: creative strategies and practical solutions, nd ed. (chicago: ala, ), . . lessig, free culture, . . ibid., . . lawrence lessig, the future of ideas: the fate of the com- mons in a connected world (new york: vintage bks., ), – , . . lessig, free culture, – . . ibid., . . ibid., – . . james boyle, shamans, software, and spleens: law and the construction of the information society (cambridge: harvard univ. pr., ), . . william w. fisher iii, promises to keep: technology, law, and the future of entertainment (stanford, calif.: stanford univ. pr., ), . . lessig, free culture, – . . litman, digital copyright, – . . ibid., – . . michael godwin, digital rights management: a guide for librarians (washington, d.c.: office for information technology policy, ala, ), , www.ala.org/ala/washoff/woissues/ copyrightb/digitalrights/drmfinal.pdf (accessed may , ). digital dystopia | bailey . ibid., – . . bill rosenblatt, bill trippe, and stephen mooney, digital rights management: business and technology (new york: m&t bks., ), – . . godwin, digital rights management: a guide for librar- ians, . . david mann, “digital rights management and people with sight loss,” indicare monitor , no. ( ), www .indicare.org/tiki-print_article.php?articleid= (accessed may , ). . julie e. cohen, “drm and privacy,” communications of the acm , no. ( ): – . . rosenblatt, trippe, and mooney, digital rights manage- ment: business and technology, . . j. alex halderman and edward w. felten, “lessons from the sony cd drm episode,” feb. , , http://itpolicy.princeton .edu/pub/sonydrm-ext.pdf (accessed may , ). . godwin, digital rights management: a guide for librarians, – . . wikipedia, “analog hole,” http://en.wikipedia.org/ wiki/analog_hole (accessed may , ). . godwin, digital rights management: a guide for librarians, – . . ibid., . . robert a. gehring, “trusted computing for digital rights management,” indicare monitor , no. ( ), www.indicare .org/tiki-read_article.php?articleid= (accessed may , ). . seth schoen, “trusted computing: promise and risk,” www.eff.org/infrastructure/trusted_computing/ _tc.php (accessed may , ). . pamela samuelson, “drm {and, or, vs.} the law,” com- munications of the acm , no. ( ): – . . declan mccullagh, “congress raises broadcast flag for audio,” cnet news.com, mar. , , http://news.com .com/congress+raises+broadcast+flag+for+audio/ - _ - .html (accessed may , ). . ibid. . danny o’brien, “a lump of coal for consumers: analog hole bill introduced,” eff deeplinks, dec. , , www.eff .org/deeplinks/archives/ .php (accessed may , ). . siva vaidhyanathan, copyrights and copywrongs: the rise of intellectual property and how it threatens creativity (new york: new york univ. pr., ), – . . lessig, the future of ideas, – . . educause, “net neutrality,” www.educause.edu/ c o n t e n t . a s p ? pa g e _ i d = & pa r e n t _ i d = & b h c p = (accessed may , ). . electronic frontier foundation, “dearaol.com coalition grows from organizations to in one week,” mar. , , www.eff.org/news/archives/ _ .php# (accessed may , ). . catherine yang, “is verizon a network hog?” business- week, feb. , , , www.businessweek.com/technology/ content/feb /tc _ .htm (accessed may , ). . ibid. . jonathan krim, “executive wants to charge for web speed,” washington post, dec. , , d , www.washingtonpost .com/wp-dyn/content/article/ / / /ar .html (accessed may , ). . harold furchtgott-roth, “at&t, or another telecom takeover,” the new york sun, mar. , . www.nysun.com/ article/ (accessed may , ). (see also: www.furchtgott -roth.com/news.php?id= (accessed may , ). . john windhausen jr., good fences make bad broadband: preserving an open internet through net neutrality (washington, d.c.: public knowledge, ), www.publicknowledge.org/ content/papers/pk-net-neutrality-whitep- (accessed may , ). . peter suber, “three gathering storms that could cause collateral damage for open access,” sparc open access news- letter, no. ( ), www.earlham.edu/~peters/ fos/newsletter/ - - .htm#collateral (accessed may , ). . j. d. lasica, darknet: hollywood’s war against the digital generation (new york: wiley, ), . . john borland, “freenet keeps file-trading flame burn- ing,” cnet news.com, oct. , , http://news.com.com/ - - .html (accessed may , ). . creative commons, “attribution . ,” http://creative commons.org/licenses/by/ . / (accessed may , ). . lawrence liang, “a guide to open content licenses.” http://pzwart.wdka.hro.nl/mdr/research/lliang/open _content_guide (accessed may , ). . peter suber, “open access overview: focusing on open access to peer-reviewed research articles and their preprints.” www.earlham.edu/~peters/fos/overview.htm (accessed may , ); charles w. bailey jr., “open access and libraries,” in mark jacobs, ed., electronic resources librarians: the human ele- ment of the digital information age (binghamton, n.y.: haworth, ), forthcoming, www.digital-scholarship.com/cwb/oa libraries.pdf (accessed may , ). . lasica, darknet, – . . waxer and baum, internet surf and turf revealed, . . lessig, free culture, – . . joe mandak, “internet archive’s value, legality debated in copyright suit,” mercury news, mar. , , www .mercurynews.com/mld/mercurynews/news/local/states/ california/northern_california/ .htm (accessed may , ). . arnold p. lutzker, primer on the digital millennium: what the digital millennium copyright act and the copyright term exten- sion act mean for the library community (washington, d.c.: ala washington office, ), www.ala.org/ala/washoff/wois sues/copyrightb/dmca/dmcaprimer.pdf (accessed may , ). the chamberlain group inc. v. skylink technologies inc. deci- sion offers some hope that authorized users of drm-protected works could legally circumvent drm for lawful purposes if they had the means to do so (see: crews, copyright law for librarians and educators: creative strategies and practical solutions, – ). continued on page toward a twenty-first-century library catalog | antelman, lynema, and pace copyright © by charles w. bailey jr. this work is licensed under the creative commons attribution- noncommercial . license. to view a copy of this license, visit http://creativecommons.org/licenses/by-nc/ . / or send a letter to creative commons, howard st., th floor, san francisco, ca, , usa. bailey continued from ฀ known-item questions . “your history professor has requested you to start your research project by looking up background information in a book titled civilizations of the ancient near east.” a. “please find this title in the library catalog.” b. “where would you go to find this book physically?” . “for your literature class, you need to read the book titled gulliver’s travels written by jonathan swift. find the call number for one copy of this book.” . “you’ve been hearing a lot about the physicist richard feynman, and you’d like to find out whether the library has any of the books that he has written.” a. “what is the title of one of his books?” b. “is there a copy of this book you could check out from d. h. hill library?” . “you have the citation for a journal article about photosynthesis, light, and plant growth. you can read the actual citation for the journal article on this sheet of paper.” alley, h., m. rieger, and j.m. affolter. “effects of developmental light level on photosynthesis and biomass production in echinacea laevigata, a federally listed endan- gered species.” natural areas journal . ( ): – . a. “using the library catalog, can you determine if the library owns this journal?” b. “do library users have access to the volume that actually contains this article (either electronically or in print)?” ฀ topical questions . “please find the titles of two books that have been written about bill gates (not books written by bill gates).” . “your cat is acting like he doesn’t feel well, and you are worried about him. please find two books that provide information specifically on cat health or caring for cats.” . “you have family who are considering a solar house. does the library have any materials about building passive solar homes?” . “can you show me how would you find the most recently published book about nuclear energy policy in the united states?” . “imagine you teach introductory spanish and you want to broaden your students’ horizons by expos- ing them to poetry in spanish. find at least one audio recording of a poet reading his or her work aloud in spanish.” . “you would like to browse the recent journal litera- ture in the field of landscape architecture. does the design library have any journals about landscape architecture?” appendix a: ncsu libraries catalog usability test tasks information technology and libraries | december digital tool making offers many challenges, involving much trial and error. developing machine learning and assistance in automated and semi-automated internet resource discovery, metadata generation, and rich-text identification provides opportunities for great discov- ery, innovation, and the potential for transformation of the library community. the areas of computer science involved, as applied to the library applications addressed, are among that discipline’s leading edges. making applied research practical and applicable, through placement within library/collection-management systems and ser- vices, involves equal parts computer scientist, research librarian, and legacy-systems archaeologist. still, the early harvest is there for us now, with a large harvest pending. data fountains and ivia, the projects discussed, dem- onstrate this. clearly, then, the present would be a good time for the library community to more proactively and significantly engage with this technology and research, to better plan for its impacts, to more proactively take up the challenges involved in its exploration, and to better and more comprehensively guide effort in this new territory. the alternative to doing this is that others will develop this territory for us, do it not as well, and sell it back to us at a premium. awareness of this technology and its current capabilities, promises, limitations, and probable major impacts needs to be generalized throughout the library management, metadata, and systems communi- ties. this article charts recent work, promising avenues for new research and development, and issues the library community needs to understand. t his article is intended to discuss data fountains (http://datafountains.ucr.edu) project work and thinking (and its foundation in the ivia system, http://ivia.ucr.edu) regarding tools and services, for use in collection creation and augmentation. both systems emphasize automated and semi-automated internet resource discovery, metadata generation, and rich-text harvest. these areas of work and research occur within the larger realms of machine assistance and machine learning. they are of critical value to libraries as they currently or potentially concern: significant resource savings; ampli- fication and re-tasking of expert effort to better match librarian expertise with tasks that truly require it (through the automation of routine tasks); and better scaling of col- lections by providing them the technological wherewithal to grow, as appropriate, and better match the explosion of significant available knowledge and information that the internet has accelerated. this article is organized into three major sections: ■ part i details machine assistance work to date in the data fountains and ivia systems project. ■ part ii describes current and upcoming promising research directions in machine assistance. ■ part iii delves into planning and organizational issues that may arise for the library community as a result of these technologies. ■ part i: recent work in data fountains and ivia part i covers work to date on data fountains and ivia. section , “a new service and open source software,” describes concrete project work with data fountains, a new open service and suite of open-source software tools for the educational and library communities, in developing practical machine learning to provide machine assistance in collection building. data fountains is an expansion of work based upon the ivia systems foundation. it is an effort that has been ongoing and evolving since . section , “role and niche definition for machine assis- tance in collection building,” covers recent developments in our ongoing effort to better research and define roles and niches for machine assistance of the types offered by data fountains. the spectrum—ranging from collection building with an emphasis on expertise that receives small assists from machine tools to an emphasis on machine tools that are configured and thereafter assisted through small refinements by expertise—is examined. results from an initial exploratory survey in these areas are summarized. ■ a new service and open-source software—data fountains description data fountains is an internet resource discovery, meta- data-generation, and selected, full-text harvesting service as well as the open source (lesser general public license machine assistance in collection building: new tools, research, issues, and reflections steve mitchell steve mitchell (smitch@ucr.edu) is the science librarian for ivia/nsdl data fountains/data fountains projects, science library, university of california, riverside. machine assistance in collection building | mitchell (lgpl) and general public license (gpl) licensed) soft- ware that makes the services possible. it is a set of tools for use by organizations and institutions serving the greater learning community that create and maintain internet por- tals, subject directories, digital libraries, virtual libraries, or library catalogs with portal-like capabilities (ipdvlcs) containing significant collections of internet resources. it is an evolved variant of the ivia system, with which it shares many components. the data fountains/ivia code base rep- resents more than , lines of primarily c++ code. on the systems level, data fountains operates as an array of independent systems containing crawler, text classifier, text extraction, portal, and database software components customized to the needs of participat- ing projects. each cooperator and subject community works with, fine tunes, and benefits from its own set of crawler(s), classifier(s), and database manager(s), i.e., its own specific data fountain. note that in this article, data fountains’ portal/metadata repository/database man- agement, content management, import-export, or content search/browse capabilities, which are substantial, will not be discussed. instead, the article will focus on its machine assistance and machine-learning components. the data fountains system and service has been developed through a research partnership among com- puter scientists and academic librarians that is beginning to provide technological solutions to some of the major overall problems associated with the scalability and effi- cient running of ipdvlcs. much project effort is based on applying machine-learning techniques to partially automate and provide help in a number of laborious and costly ipdvlc activities. included here, more specifically, are the following needs/scaling challenges: reducing to some degree the high costs of manually created metadata; better coverage of the ever-increasing number of important internet resources (relatedly, the relatively small size of most library internet collections, where searches yielding very few or no results are common); reducing or making more efficient expert-involved tasks requiring little exper- tise; and reducing redundant efforts among ipdvlcs (both in content and systems building). by providing inexpensive, universally needed raw materials (i.e., metadata and rich full text represent- ing important resources), the data fountains service is intended to offer major support and resource savings to cooperating ipdvlc participants that otherwise have strong ongoing commitments to their established institu- tional identity or “brand,” interface or look, system, and, more generally, “established way of doing things.” data fountains viability and sustainability is keyed to providing universally needed service and very generic information products that do not require ipdvlcs to change—this often being seen as prohibitively expensive in time and resources. data fountains is intended to lower barriers for substantive cooperation in collection building and resource savings on the part of large numbers of ipdvlcs by developing, sharing, and distributing the benefits of machine learning in its areas of application. the data fountains service will be useful to a large spectrum of academic and library-based finding tools including metadata repositories and catalogs with internet portal-like capabilities. increasingly, library-catalog soft- ware is developing more flexibility, including, hopefully, the means by which full marc (machine-readable cataloging) records coexist with more streamlined (and less expensive) records, e.g., dublin core (dc) and other types, and, moreover, metadata records that include or can be closely associated with selected rich full-text, among many other catalog need areas. data fountains offers mul- tiple levels of products and services geared to fit the needs of ipdvlcs of differing sizes, subject needs, and desired data “completeness” or depth (this being the amount and type of metadata and full-text needed to properly represent each resource). uses, products, and services overall, data fountains automatically or semi-automati- cally supplies varying levels of what represents the basic “ore” required by ipdvlcs for internet resource and article collection building: access to significant, previ- ously undiscovered resources as well as the metadata and selected full-text that describe or represent them. this ore is available in both raw (relatively unprocessed) and more refined products depending on the needs of the partici- pating ipdvlc including, perhaps most importantly, the degree to which expertise is available to provide further refinement and how and for whom the material is intended to be used. data fountains multiple product and usage models supports the building of a wide array of ipdvlc collections. a number of usage or service models are supported by data fountains, including: collection development support for single hybrid record type collections the first usage model, based on full automation, involves the utilization of data fountains metadata and rich, full- text “as is,” without review, to populate a collection. these records can be used by themselves or mixed with other types of records. they can also be used as part of a hybrid collection to undergird another, more primary, or fully expert-created, collection. while more accurate, expert- created collections are not only comparatively more labor intensive and expensive to create and maintain, but often smaller, with narrower and more limited coverage. this has been the infomine (http://infomine.ucr.edu) model that features two distinct collections, with the automati- cally generated collection supporting, as a second tier of information technology and libraries | december data, the expert-built content in the primary collection. users can search one or both. internet resource discovery service a second model uses data fountains primarily as an internet resource discovery service where links and titles and other minimal metadata are supplied but where the user’s intent is to identify new resources and build metadata records emphasizing a considerable amount of metadata not generated by data fountains (e.g., different subject schema). this is done by utilizing the targeted link crawler, expert guided crawler, or focused crawler. because little to no metadata/rich-text generation/extrac- tion occurs, this is the least complex of the usage models. crème de la data fountains a third approach, a variation of the second, utilizes only those data fountains records that have been automatically determined, through a user-set threshold, to represent the most highly significant resources (e.g., the top percent). these can be flagged for expert review or automatically harvested without review. the data fountains metadata retained for expert review, post-processing, and improve- ment can be minimal or full. metadata records intended for expert refinement a fourth approach, which is semi-automated, involves using data fountains as both a discovery service and as a metadata record-building service where employment of records from the data fountains data stream is selective but the machine-created record is routinely retained as a foundation record to be refined or augmented by the expert. metadata records plus full-text a fifth approach is to use the rich full-text selectively iden- tified and harvested from the internet resource, either in addition to the metadata generated or by itself, to populate a collection and greatly boost retrieval. that is, some col- lections may want to utilize metadata differing from that produced by data fountains but have data fountains perform the service of augmenting their metadata with rich full-text. all or parts of the object and full-text can be harvested. ■ standards, metadata, and full-text data fountains’ record format is dublin core (dc) and features standard research library subject schemas includ- ing slightly modified library of congress subject headings (lcsh) and library of congress classification (lcc). as part of upcoming work, development of additional classifiers to apply other subject/classification schemas/ vocabularies will occur, notably ddc and those that can be automatically invoked from the terminology found in the collection objects. cooperators may choose to help develop new formats, subject schemas, and metadata to meet custom needs in collecting and classification. other important metadata generated include: title, creators, description (an annotation-like construct), keyphrases, capitalized terms, and resource language, among a total of thirty-plus fields. in addition to fielded metadata, data fountains delivers selected rich text harvested from the resource. this is important for enhancing ipdvlc retrieval capabilities and user-searching success. the rich text can be harvested verbatim and offered as-is for search or, if this is problematical, further processed into keyphrases. data post-processing, transfer, and product relevance assurance participants determine and download resources of rel- evance automatically in batch mode via subject-profiled, custom internet crawls and editable results sets created by and for each ipdvlc to reflect its particular interests. these profiled crawls and metadata generation routines are stored and can be re-executed at selected intervals. results are transferred using the open archives initiative protocol for metadata harvesting (oai-pmh) or sdf (standard delimited format) in dc, marc, and extensible hypertext markup language (xhtml) formats. in addition to batch transfers, participants can manually and interactively identify individual records or groupings of records that suit their needs for harvest. selective, interactive, sort- ing/browsing of results, followed often by evaluation and editing of metadata and full-text fields (as individual records or globally in patterns), is enabled prior to export. these capabilities allow precisely targeted, custom record identification, modification, and downloading. this in turn enables the most general, as well as the most subject-spe- cialized, ipdvlcs certainty in identifying and receiving only records that meet their need criteria. open-source software the software making the above possible is available to all for free through the lgpl/gpl open-source licenses and model. the open-source model should work well for tool development as fundamental as that described. open source of this type generally means that users freely use and perhaps participate in further development of the functionality of the software and, at intervals, contribute their innovations back to the code base for all to use. lgpl/gpl supports a wide diversity of forms of com- machine assistance in collection building | mitchell mercial service development. open source has worked well for large applications such as many forms of the linux operating system (a number of variants of this are supported), apache server software, and mysql database management software (all of which are used by the data fountains system). using this model has the intent of cooperatively benefiting the community as a whole. it is the author’s belief that tools of the data fountains type will have wide enough usage within and are crucial enough to the library community to support the development of an open-source community around them. data fountains software is of use to thousands of institutions that build ipdvlc collections. open source also means that the development and evolution of a core tool or system for a community can potentially occur faster and more flexibly, with the proper community support, than many types of proprietary effort. this is needed given the continuing and increasingly greater revolutions in computing power and software potential. the community needs to be able to evolve faster in response to changing conditions, and free, community- based, open-source software development is one strategy for achieving this. ■ current systems design, development, and features to date, most of the work has emphasized research and development leading to innovations in preferential focused crawling, subject classification using logistic regression, knearest neighbor (knn) and other classifiers, and rich full-text identification and extraction. a major emphasis in systems development has been identifying points of intervention in crawling, classification, and extraction, whereby initial, periodic or ongoing interactive expert input can be employed to improve machine processes and results. that is, the work has emphasized usage not only of fully automated machine processes but semi-automated machine processes intended to interactively augment, amplify, and improve the efforts of experts. experts assist machine processes, and machine processes assist expert judgment/labor. the programming has also been done with an eye toward modularity among different systems components. ■ internet resource discovery/ identification—expert guided and focused crawling a number of crawling systems have been used; cur- rently, for data fountains, three are used that represent two approaches to crawling: expert guided and focused. expert-guided crawling is accomplished by a targeted link crawler (tlc) and an expert guided crawler (egc). tlc is concerned with crawling a user-specified link or list of links. egc differs from tlc in that the single “start url” link given is only the beginning point from which the crawler will either drill down (find onsite links at multiple depths in a site) or drill out (find external links not on the start url site). the result is that, compared with tlc, many more links than just those given the egc crawler initially are crawled. with all crawlers, a metadata record with accompanying rich full-text is generated for each resource crawled. a preferential focused crawler, called the nalanda ivia focused crawler (nifc) after the name of the ancient seat of learning in india, continues to be developed. focused crawling makes possible focused identification of signifi- cant internet resources by identifying specific, interlinked, and semantically similar communities of sites of shared subject interest. generally, nifc traverses subject expert- targeted regions of the internet to find resources that are strongly interlinked and thereby represent coherent subject-interest communities and sites of shared interest and mutual use (i.e., are often concerned with and contain content similar to one another). communities sharing interests often identify and cite one another through link- ages on their internet resources. through this mechanism, these communities and their sites/resources can be identi- fied, mapped, and harvested. preferential focused crawl- ing makes focused crawling more efficient by employing algorithms that can respond to clues in web resource page layout and structure (e.g., using document object models, visual cues, and text windows adjacent to anchor text, among others) that indicate the more “promising” links to crawl. the result is more efficient focused crawling (figure ). the focused crawling process starts with exemplary sites/pages/urls being supplied by participating ipdvlc experts. these highly on-topic exemplars are used to form a seed set of model pages used for training/guid- ing the crawler. as the crawling progresses, an interlinkage graph is developed of which resources link to one another (i.e., cite and co-cite). highly interlinked resources are evaluated, differentiated, and rated as to the degree to which they are linked to/from as well as for their capaci- ties as authoritative resources (e.g., a primary resource such as an important technical report that receives many in-links to it from other resources) or hubs (e.g., secondary sources such as expert virtual library collections that pro- vide out-links to other, authoritative resources). as hubs, expert-created, high-quality ipdvlc collections of links (e.g., infomine) play an important role as milestones and navigation aids in the guidance of many types of crawling. another automated process works to rate resources, as a second indirect measure of resource quality, by comparing for similarity of content (e.g., similarities among key-word information technology and libraries | december vocabularies) between the potential new resources and model resources. the most linked to/from authorities and hubs, with terminology most similar to the exemplars, are thus identified and become prime candidates for adding to the collection and for indicating other resources to add. the overall architecture of data fountains involves multiple concurrent crawls and an array of multiple crawlers and associated classifiers on multiple machines (i.e., there are one or more data fountains for each major subject area or major cooperator). areas of expert interaction in focused crawling expert interactive and semi-automated approaches to improve crawling are employed in and constitute special design areas of data fountains since many participating projects and communities have access to considerable subject expertise. there is much promise in amplifying the role of this expertise in the crawling process. experts can create and refine crawls by: ■ determining the most appropriate seeds (exemplary resources) to use (whether found in their own collec- tions or generated from other sources); ■ choosing degree of “on-topic-ness” desired (a preci- sion versus recall setting); ■ determining the total number of resources to be crawled; ■ editing initial crawl results (e.g., de-selecting or blacklisting resources found) with an eye toward generally refining and developing a super seed set of very large numbers of increasingly on-target seeds that are then crawled anew. (this process of refine- ment and enlargement can be reiterated as desired in achieving increasing accuracy in and numbers of exemplars and therefore accuracy in the final crawl.) ■ in addition, expert truing of crawler web graph weightings (i.e., manually “lifting” the values of selected hubs and authorities) either during or after a crawling run is being explored to improve crawling accuracy. this lifting process can be aided through tools to visualize the crawl so that the expert can quickly identify, among the masses of results, the most promising areas of a web graph for the crawler to emphasize. ■ expert-created blacklists of urls for types of sites or pages that are not valuable can be stored to save future crawling and expert time. there is such a blacklist for each participating data fountains com- munity group and individual. ■ metadata generation— automated and semi-automated subject classification data fountains and ivia embody innovations in automated metadata generation, including identifying and applying controlled subject terms (using academic library-stan- dard subject schema), keyphrases, and annotation-like constructs (figure ). automated classifier programs apply these and other metadata and are part of a suite of programs known as the record builder. controlled subject terminology applied currently includes lcsh, lcc, ddc, and medical subject headings (mesh). in assigning these, the system generally first looks for html and dc metatags and then extracts these data. with some fields, when these data are not present (which is common), original metadata are then generated automatically. in the case of lcsh, lcc, and ddc, if not present in metatags, or if users choose to override metatag extraction (in cases where metatags are not accurate, such as when they are spammy or when top-page boilerplate metadata is carried onto all pages regardless of subject relevance), then classification processes are invoked. these derive a set of keywords and key phrases from the resource that serve as a surrogate in representing and summarizing its content. then, using a model that encapsulates the rela- tionships between these natural-language terms and the set of controlled-subject terms, the closest corresponding set of controlled terms is assigned. the model is learned from training data sets that consist of large sets of records (more than thirty million in corpora loaned for research purposes by the cornell university library, library of congress, california digital library [cdl], and oclc) figure . focused and preferential crawling (courtesy of s. chakrabarti) machine assistance in collection building | mitchell from library catalogs and virtual libraries. with lcc, the aim has been to assign one or more lccs to a resource based on the set of lcshs associated with that resource. svm, knn, and logistic regression classifiers have been used. generally, performance has been acceptable in cases where there were two hundred examples of the usage of a particular lcsh (in a record with a url). unfortunately, as large as the training data sets have been, there simply haven’t been enough records for classification purposes with urls and associated text. this problem will more than likely be resolved shortly as catalogs increasingly incorporate web resources. metadata generation—automated extraction of known, named entities named-entity (e.g., data elements that can be expected to be in a resource and that are placed by authors/publish- ers within a known textual/markup pattern) extraction is primarily practiced through the simple means of identify- ing and extracting data elements indicated by html/dc metatags, when present on a page. data for more than thirty dublin core common (and not so common) fields are extracted. with some fields, extraction can be guided, as needed, in the interests of original metadata creation through pattern recognition and profiling, or through classification (e.g., title, subjects, description). ■ rich-text identification and harvest refinement of our “aboutness” measure for identifying the most relevant pages or sections in a resource or document (i.e., those intended by the author to be rich in descrip- tive information about the topics within and the type of resource) from which to extract text is a continuing pursuit. involved in this quest has been better determination of author-created structures and conventions in document or resource layout (e.g., locating introductions, summaries, etc., and determining/proportioning the amount of text to be extracted from each). more accurate rich-text identification in turn yields more accurate identification, extraction, and application of key phrases and, from these, more accurate controlled subject term and other metadata application. this is at the foundation of many metadata generation processes. crucially, rich full-text is also important from an end-user information-retrieval perspective because the natural-lan- guage terminology contained partially corrects for the limi- tations inherent in many controlled metadata and subject vocabulary/schema approaches (e.g., new or specialized subject terminology is often slow to appear or weakly represented in the often generalist library-standard subject schemas). refinement of the “aboutness” measure in identi- fying terms indicating that rich text follows is an important and ongoing task that involves formulating fairly intricate text-extraction rules in reflecting conventions in rich-text placement in resources and documents of differing types (e.g., web sites, articles, database interfaces), formats (e.g., html, pdf, postscript), and languages. ■ a modular architecture that supports a federated array of subject-specific focused crawlers and classifiers the architecture that data fountains is based upon is shown in figures and . data fountains operates on the systems level as an array of separate sets of bundled crawl- ers (both guided and focused), classifiers, and extractors; this bundled array of crawlers approach provides greater flexibility and efficiency, as compared with using a more monolithic, single-crawler, multiple-subject approach. a bundle can occupy a whole machine or several can exist independently, as virtual data fountains, on a single machine. instead of one broad, multiple-subject, multiple- audience data fountain that follows a broad shotgun approach to internet resource discovery and classification, there are several vertical, subject- and audience-focused data fountains. a data fountain is intended to exist for each distinct, major subject area and the subject-specific ipdvlc collections (e.g., visual arts, business, horticulture) associated with them. figure . metadata choices in data fountains information technology and libraries | december data fountains systems architecture emphasizes modularity. it has been enabled and assumed that sepa- rate components of the system (e.g., the crawlers, classi- fiers, database management systems) could be developed further for other uses independent of the data fountains system. in addition, as technologies that the system is dependent upon advance, users will be able to more easily swap out and replace older modules. these capabilities contribute to system sustainability. ■ service design and sustainability data fountains was conceived to be a cooperative, non- profit, low-overhead, cost-recovery-based service intended to sustain itself after start-up. access will be provided to ipdvlc cooperators who demonstrate interest and sup- port for the work and service. by so doing, cooperators share in supporting the continuing evolution and improve- ment of data fountains. as an additional sustainability consideration, the software has been released as open source so that it can develop and evolve in many directions (to directly fit unique needs) as well as benefit through distributed effort. ■ “small is beautiful”: roles for and advantages of appropriate small- to medium-scaled tools approaches like those data fountains has taken may be among the few ways that internet finding tools can con- tinue to be relevant to the learning/library community and offer the accuracy and significant content needed by that community. the technical challenges faced by the large engines in their quest to cover an infinitude of audiences and internet resources do not need to be grappled with by the community of research libraries and are not faced by focused crawlers and classifiers of the type data fountains relies upon. the latter are better able to develop targeted, more accurate approaches to their subjects because they enable machine assistance for, as well as amplification of, authoritative subject expertise (e.g., librarians) as a core interactive component in the process of finding and describing new resources. the processes involved target more narrowly defined, distinct, and finite subject uni- verses and intellectual communities. this, in turn, allows them to scale appropriately for their tasks and to apply more complex and varied types of metadata for faculty, researchers, graduate students, and librarians, who gener- ally require more precision (and authority) in their finding tools but still need to move beyond collections (even allied) that are essentially catalogs moved forward a notch. the smaller scale of this work also potentially enables inno- vations in effective linkage and similarity (i.e., semantic) analysis. some experts note that the future of internet searching as a whole may lie in searching federated finding tools based in these techniques. such a federation could be an academic’s or librarian’s web of high-quality find- ing tools. data fountains may offer part of the foundation needed to support such a web. from a related perspective, these tools represent an appropriate approach for library and library-community- scaled resource identification and description tasks that emphasizes perhaps the great advantage the library com- figure . interaction of fully and semi-automated and manual collection building processes in data fountains figure . overall data fountains architecture machine assistance in collection building | mitchell munity can bring to bear in creating useful finding and metadata generation tools, which no others have. that is, the community’s unparalleled subject and description expertise in finding and ordering significant resources into coherent rich collections might be amplifiable shortly, through machine assistance. if such an effort was sensi- bly coordinated and focused, and minor modifications in approach and established standards made to enable best use of these new tools, then the best internet find- ing tools/collections could be made possible yielding high-quality and significant coverage. these collections would benefit by having the capability to catalyze, out of the mass of the web, the resources that constitute much of its intelligent fraction and make this coherently visible and available to learners and researchers. moreover, this could be done in such a way that digital and print record and object collections could seamlessly interact as one, rendering what would be the best information-finding tools/collections without regard to type of resource. this effort in fact has been unfurling for a long time, though, to date, in small and somewhat sporadic, uncoordinated ways. for example, infomine and similar collections have provided credible links to and for the academic com- munity for well over a decade. ■ role and niche definition for machine assistance in collection- building exploratory survey an exploratory survey conducted in fall illuminated new perspectives, desired products and services, and research opportunities as perceived by a sampling of digital library and library leaders in regard to a number of areas involving machine assistance in collection building. generally, areas explored concerned, among others: new roles projected for machine learning/machine assistance in libraries for metadata generation, resource discovery, and rich full-text identification and extraction; new finding-tool niches and opportunities existing in the service spectrum between google and opac; acceptance of streamlined, more minimal, and cost-saving approaches to metadata creation or augmentation; the role of cost-recovery-based service and cooperative, participatory business models in digital libraries. more specifically, the purposes of the survey were to: . elicit leading library attitudes in relation to the types of services, software development, and research that generally will constitute data fountains; . test the waters in regard to attitudes toward implementing machine-learning/machine-assis- tance-based services for semi-automated collection building within the general context of libraries; . probe for new avenues or niches for these services and tools in distinction to both traditional library services/tools and large web search engines; . concretely define data foundations’ initial set of automatically and semi-automatically generated metadata/resource-discovery products, formats, and services; . examine attitudes toward the value and roles of rich, full-text in library-related finding tools; . examine attitudes toward hybrid databases contain- ing heterogeneous records (e.g., multiple formats, types, and amounts of metadata); . gather ideas on cooperatively organizing such ser- vices; and . generally define new ideas in all interest areas for development of products and services. the survey, comprised of fifty-nine questions, was sent to thirty-five managers of leading digital libraries/librar- ies/information projects. there was roughly a percent return from those targeted (fourteen out of thirty-five). responding institutions and individuals were guaranteed anonymity of response. ■ survey result summary there was considerable agreement on most answers. as such, this initial definitional survey has proven helpful in design and product definition. though the survey sample set/number of respondents was limited and while results need to be seen as tentative, the views expressed are from well-regarded experts in the fields of digital library and library technology, development, and services. in addi- tion to helping define current data fountains services, the survey results also indicated the need for further explora- tion in the areas of services, tools, overall niche definition, and publicity. while conclusions remain tentative, barring future, larger surveys, some of the more relevant results are as follows: ■ there appear to be significant niches for an auto- mated/semi-automated collection-building/aug- mentation service given inadequacies in serving research-library users found in google (and presum- ably other large commercial search engines) and commercial-library opac/catalog systems. survey results indicate a need for services of the types char- acterized by data fountains. ■ generally, academic libraries get a slightly above middle-value (neutral) grade in terms of meeting shifting researcher and student information needs over the last decade. this indicates that, above and beyond specific library and commercial-finding tools, information technology and libraries | december there are information needs not being met by librar- ies in regard to information discovery and retrieval that new services may be able to help provide. ■ there is support, above and beyond creating machine-assistance-based collection-building ser- vices, for developing and distributing the free, open- source software tools supporting these services. tools that make possible machine assistance in resource description and collection development are seen as potentially providing very useful services. ■ automated metadata creation and automated resource discovery/identification, specifically, are perceived as potentially important services of signifi- cant value to libraries/digital libraries. ■ there is support for the notion of automated iden- tification and extraction of rich, full-text data (e.g., abstracts, introductions) as an important service and augmentation to metadata in improving user retrieval. ■ the notion of hybrid databases/collections (such as infomine) containing heterogeneous metadata records (referring to differing amounts, types, and origins of metadata) representing heterogeneous information objects/resources, of different types and levels of core importance, was supported in most regards. ■ many notions that were foreign to library and even leading-edge digital library managers/leaders (the respondents) two to three years ago appear to be acknowledged research and service issues now. included among these are: machine assistance in collection building; crawling, extraction, and clas- sification tools; more streamlined types of metadata; open-source software for libraries; limitations of google for academic or research uses; limitations of commercial-library opac/catalog systems; and the value of rich full-text as a complement to metadata for improved retrieval. ■ there is strong support, given the resource savings and collection growth made possible, for the notion of machine-created metadata: both that which is cre- ated fully automatically and, with even more sup- port, that which is automatically created and then expert reviewed and refined. ■ amounts, types, and formats of desired metadata (very streamlined dc metadata was supported for most uses and contexts) and means of data transfer (oai-pmh was preferred) were specified by respondents. ■ summary of part i data fountains is a unique service and system for inex- pensively supporting aspects of collection building among ipdvlcs. developing and utilizing advances in focused crawling and classification, this service automatically and semi-automatically identifies useful internet resources (both open-web as well as closed-collection resources including articles and reports, etc.) and generates metadata (and selected rich text) to accompany them. data fountains is a cooperative service, a free open-source software sys- tem, and a research-and-development project exploring machine assistance as well as machine-expert interfaces and synergies in collection building. several useful service niches and roles for the work have been identified and have been or are being developed. ■ part ii: new directions in research this section discusses important new directions in research for machine assistance in collection building as they relate to upcoming and expanding research, development, and prototyping within data fountains and ivia. among focus areas are promising means of: automated classification for applying library standard controlled subject vocabular- ies/schema, including hybrid and ensemble classification; smarter and more accurate named-entity extraction (e.g., capturing object/article metadata “facts” such as publisher and publishing date); improvements in rich-text identifi- cation and harvesting; article/report collection level co- citation and subject gisting functionality; and generally improved expert-guided and focused web crawling. ■ new research in machine assistance for collection building the ivia and data fountains projects have recently received a fourth national leadership grant from the united states institute of museum and library services that supports three years of research and development in machine assistance in collection building. in addition, the national science digital library is continuing funding. the areas that will be worked in are discussed below. these have been determined through experience gained over the last eight years of work in machine-assistance-oriented systems development and dialogue with computer scientists and collection coordinators. these areas of technology work and application, though complex and challenging, are very important. that is, assuming it is important that the learn- ing/library community not be dis-intermediated by such technologies but instead becomes more fully empowered by them. this can only occur through developing a much larger role in actively defining, guiding, and putting the technologies to best possible use. looking into the future, it is clear that libraries cannot simply continue to wait for or rely on good companies like machine assistance in collection building | mitchell google, oclc, or opac creators to deliver them, much like a cargo cult, as they have in the past. to the degree that this is done, there is the risk of becoming vendor vectors blinded by the limitations of these companies and their product lines. these products are often incorrectly assumed to be the known technical and organizational universe of what is possible or doable. the revolutions coming in computing power together with the low cost of this power—which will be almost ubiquitously distributed among users of library collections and services—promise much more change than libraries have seen in the last decade. among the changes underway are those in machine learning and machine assistance in libraries. as the changes take place, organization size may not guarantee much as, over the last decade, librarians and researches have witnessed large academic and other research libraries, with some exception, demonstrate a profound organizational entropy in almost direct propor- tion to the magnitude of what are essentially paradigm shifts in scholarly communications, information provi- sion, and research. to some degree, these simply reflect larger blockages within the universities and institutions in which libraries are embedded. as these changes play out, it should be noted that history in information or library- related public or scholarly information provision/access probably will not end with google or oclc—wonderful and fairly open companies—just as history in automobile manufacture has not ended with gm, computer manufac- ture with ibm, or web finding with alta vista. with this as background and in the vein of open planning (as well as open services and open software) and given the size of the work areas addressed and their challenges, much of the projects’ technical planning and direction are being presented in this paper. these areas of computer and information-science research and develop- ment, which will affect libraries in many ways, are evolving rapidly into practical application. the current major research areas are: ■ named-entity identification and extraction, and unified models of information extraction and data mining named-entity identification and extraction is concerned with finding and harvesting generally concise factual data—often common bibliographic metadata—present in the targeted resource such as publisher, title, and publish- ing date. this type of metadata usually is associated with particular collections containing information objects that are often homogeneous (e.g., scientific article collections) and in which author-intended placement of metadata (or data) elements follows an established pattern and location in the object (e.g., an abstract is typically present and indi- cated in a pattern following presentation of title/author). while making extraction easy is one of the functions of metatagged metadata in internet resources, generally few authors or collection coordinators in academia, or else- where, use metatags or applicable naming schema in any significant or uniform way (often, in fact, it is used very sparingly or not at all). extractors therefore must be able not only to identify and harvest metatag metadata, but must discern and then extract specific metadata elements interspersed in bodies of text, as made identifiable by detecting the patterns of occurrence unique to the type of element as it occurs in the object or collection. among the many advances planned for data foundations is the usage of conditional random fields. important as well are user interfaces or dash boards that allow configuration of extractors whereby, as patterns of placement for desired data for extraction change in differing collections and types of objects, the tool can be configured appropriately to match the context and task. also under development consideration are more hybrid, unified approaches to and models for data extraction and mining (as applies to text classification), using each to inform and improve the other. that is, a family of models is being developed for improving data mining of informa- tion in largely unstructured text by using methods that “have such tight integration that the boundaries between them disappear, and they can be accurately described as a unified framework for extraction and mining.” much of this work is concerned with generating metadata for article/report-level collections. ■ document-scale learning and classification a strong emphasis in the new work will be on document- scale machine learning, classification, and named-entity extraction in regard to collections of research papers, reports, theses, and monographs. internet-object boundary detection is another impor- tant concern. detecting and properly defining compound documents (e.g., web hyper-books on multiple pages or sites) is a goal, as is identifying compound-document points of author-intended entry and intended-user paths (i.e., author-intended main connective threads in distrib- uted or compound documents). relatedly, improved internal-document structure identification for better docu- ment-level classification and extraction is critical. involved are standard-document internal-structure identification (e.g., abstract, introduction, summary text, captions for tables/figures) including units of rich text and micro- information units of text organized via subtopic. methods of document-level word-and-phrase graphing as per information technology and libraries | december textrank and other means of identifying small-world and micro-information units are currently being pursued. a strong emphasis as well will be on examining and implementing new means of co-referencing among docu- ments in collections and new means of identifying latent topics in a well-defined collection. by way of explanation, another term for co-referencing is co-citation. an example of such co-referencing is referencing work, described in papers, that has been funded through the same agency and program or that shares principal investigators in addition to standard bibliographic citation. this will improve on work done in citeseer.ist (researchindex) and similar projects through integrating and advancing the promising approaches of rexa open-source collection-management software. the focus of this effort will be on integrating article-level named-entity extraction as well as co-citation and bibliometric-refined subject identification within col- lections of papers/reports. ■ individual text-classification algorithm and training method improvement new research on individual text-classification algorithms will be examined and applied. the emphasis here will be on prototyping and measuring how applicable recent promising scholarly work might be to library-related meta- data-generation challenges. the major focus continues to be in the area of applying controlled, library standard sub- ject vocabularies (e.g., lcsh, lcc, and ddc). many of the improvements relate to advances in individual text-clas- sification algorithms, classifier training and fine-tuning, training-corpora cleanup and normalization techniques, and creating the ability for the individual classifiers to hybridize with other classifiers. of special interest are classifiers that perform well with very large numbers of classes, both small and large amounts of text, and that yield probabilistic estimates in class assignment (e.g., of a particular lcsh). the latter allows both provision of multiple class assignments for resources that have mul- tiple subjects as well as greater accuracy and knowledge of the confidence level of the assignments (thresholds of confidence level in accuracy can be set in applying, or not, a particular classification). more specifically, this work will examine, test, and— depending on test results—refine recently improved variants of the most promising of several classification algorithms. among those are: ■ support vector machines (svms) ■ logistic regression (lr) ■ naïve bayes (nb) ■ hidden markov models (hmms) ■ knn/knn model a number of metrics to measure performance of these and other text classifiers in regard to controlled subject assignment, in both fully-automated and user-interactive (semi-automated) modes, will also be employed. ■ hybrid classifiers an important effort will be to test and develop new hybrid classifiers that incorporate the best capabilities of two or more in one classifier. much of the current research has involved developing and improving new hybrids that combine the best of discriminative (e.g., lr, svms, decision trees) and generative (e.g., nb and expectation maximization) techniques in classification. for example, nb is fast but lacking in accuracy, while svms are accurate but can be slow to train. hybrid models can produce better accuracy/coverage than either their purely generative or purely discriminative counterparts. various combina- tions, among others, of lr, hmm, and svm are among the most promising. ■ ensemble classification or classifier fusion this constitutes one of the main current directions in clas- sification research and has been applied to a wide range of real-world challenges. classification ensembles are reputed to be more accurate than any individual classifier making them up. an important focus is on experimenting with new approaches to automated and semi-automated ensemble classification that involves creating frameworks that support metaclassifiers or classifier-recommender systems to apply multiple classifiers, as appropriate, to the classification task. developing classifier ensembles, including the metaclassifiers to guide them, is a major element in making possible the self-service aspect of an open, automated metadata-generation service, given that the metaclassifier is intended to determine the nature of the collection and classification task and assign the appro- priate classifier(s) to the job. it is probable that expert interaction at suitable points in this process will improve performance. ■ distributed classification classifier ensembles are often used for distributed data mining in order to discover knowledge from inherently distributed and heterogeneous information sources and to scale-up learning to very large databases (often the context for library-related tasks). however, standard methods machine assistance in collection building | mitchell of combining multiple classifiers, such as stacking, can have high performance costs. new classifier combination strategies and methods of distributed interaction will be examined to better handle very large classification needs. distributed classification, by nature, would be focused on improving large-scale self-service classification. ■ semi-automated, expert-interactive classification means of enabling semi-automated, expert-interactive classification will be presented. there is much scope for building interactive classifiers that engage the tool user or collection coordinator in an active dialogue (e.g., multiple iterations of machine/expert actions and feedback loops) that leads to incorporation of expert knowledge about specific classification tasks, metadata, and collections into the classifier, thus improving performance. that is, an active learning model can be extended significantly for these processes to include both feature-selection and docu- ment-labeling conversations, exploiting rapidly increasing computing power to give the user immediate feedback on choices to improve the classification process. several different models featuring domain expert- interactive classification and extraction will be evaluated. these vary from being extremely interactive, emphasizing frequent machine assists, to less interactive, where experts profile, launch, and only occasionally refine a primarily machine process. the initial focus will be on the latter models. note that ivia and data fountains have included a metadata generator with semi-automated record builder for years. ollie and hiclass are examples of systems that are more intensively expert-interactive. classification tasks and collection types will be characterized as to which lend themselves to frequent expert interactions, occasional interactions, or more fully-automated modes (i.e., little interaction or initial profiling/definition only). ■ classifier training and evaluation techniques as important as direct work on the classifiers is work emphasizing assessment, cleaning, and testing of clas- sifier-training data and classifier-evaluation techniques. involved are training data/corpora-normalization tech- niques, document-clustering techniques, and classifier bias/variance-reduction techniques. also involved on the classifier side are tuning issues in regard to the data at hand, including improved feature-selection techniques and determining and using confidence estimates in apply- ing/not applying classifications. different approaches to these will be examined, tested, and refined with a range of training corpora. diverse training and test data from assorted collection “types” will include standardized corpora as well as data from participating library or educational community proj- ects. that is, the techniques will be assessed with regard to how they perform with: ( ) open web resources, ( ) col- lections of research papers, reports, theses, or monographs (working with rexa), ( ) typical campus web-site pages, and ( ) mixes of the above. each collection-type focus will require differing approaches, algorithms, training, and fine-tuning techniques and will be evaluated through a number of measures. ■ improved rich-text identification and extraction for improved classification and user search/ browse rich text is text that has the role of conveying through traditional or new document structures or conventions (e.g., introductions, tables of contents, faqs, and captions for figures) the author-intended subject(s) and intent of the information object. being able to accurately identify and extract this material greatly aids in classifier performance by improving significant keyphrase identification as well as in user retrieval by enabling full-text retrieval. the avail- ability of natural-language text for searching is one means of helping to resolve problems encountered in searching controlled, library standard subject vocabularies (which in turn counteract problems searchers have when only natural-language retrieval is available). both approaches are inherently complementary. improvements in rich-text identification and harvest through improved means of document-structure learning (e.g., identifying text windows around links or captions for figures and tables) will be sought. the lightweight semantic (e.g., use of terms that indicate “aboutness” such as “about”, faqs, introduction, and abstract; rating the frequency and uniformity of application of these terms in a given collection; and proportioning source of harvest) and markup clues will be refined as well. identifying aboutness text, which can be seen as micro-information units of text organized via topic and subtopic, is being pursued through work with rexa and others. ■ improved focused crawling focused crawling is an appropriately scaled method of crawling for many library collections (see part i). it is used to discover new internet resources by defined topic termi- nology and topic web-link neighborhood. topic similarity information technology and libraries | december and semantic analysis are key measures of significance that are combined with linkage co-citation measures to indicate significance or relevance of a new resource. topic similarity among resources will be increasingly modeled through a topic-linkage matrix (i.e., semantic similarity map). new means of evaluating, fine-tuning, and improving basic crawling will be examined. rules reflecting the specific semantics of each major subject area are to be developed by participants for crawls/classification. ■ combined mining and extraction that support improved focused crawling in regard to best link pursuit and expert interaction the development of hybrid, unified approaches to extrac- tion and mining can be applied to focused crawling. the processes of data mining, rich-text identification and extraction, and the newest forms of focused crawling are starting to overlap and depend upon one another in important ways (as discussed in the section on preferential focused crawling). another focus for development efforts will therefore be work to more systematically refine best- link pursuit with an eye toward combining advances in mining, extraction, and rich-text identification in focused crawling. this work will be undertaken to improve the work on nifc. focused crawling will improve in many situations, as well, through use of user-interactive compo- nents and data-visualization interfaces (e.g., control boards that visualize an interactive graph to aid in expert “lifting” of the values of specific sites/subtopic neighborhoods to better reflect their significance to the expert). this in turn that will help users guide and tune the crawling, in semi- automated fashion, to better fit the goals and context of a particular crawl. ■ modeling different approaches for a self-service, openly accessible metadata-generation service(s) the data fountains and ivia efforts have some experience with modeling metadata-collection related services, having provided collaborative, scholarly virtual-library service successfully for more than a decade. the data fountains project has improved upon earlier work and represents an automated and semi-automated resource discovery, metadata generation, and rich-text identification and harvest service for cooperating collections. the intent is that data fountains be a self-service operation. in related effort, with co-operators at the national science digital library (nsdl) and library of congress (lc), the data fountains project has been striving to develop self-service dash boards that collection managers can use to configure, profile, and satisfy their needs. by complementing initial profiling with ongoing interactive dialogue, guidance, and refinement, more precise task definition and tool utilization can be achieved. the goal is to have a service that can, through advanced interfaces, engage users in dialogue to help them better determine their options, the tasks involved in achieving them, the capabilities and limitations of the tools available, and therefore, the best choice of tools and practices given their specific service needs and the nature of their collections. ■ summary of part ii there are many fronts of research in machine learning as applied to text processing and new-resource discovery in regard to collection building of various types, relevant to libraries, which have opened over the last few years. the data fountains/ivia research described is looking into just a few of these. for libraries, the borders between computer science, information science, and library science are dis- solving rapidly. it would be hard to devise or project for- ward a five-year plan for a large working library without some understanding of current and oncoming machine- learning and machine-assistance work in each of these disciplines, the many inter-connected organizational/com- munity/technical issues, and without an understanding that goes beyond the domain of current or developing products and services from existing vendors. ■ part iii: issues and reflections part iii is intended to define and address some of the many challenges and issues that are arising or may arise as a result of work on machine-assistance tools in the areas of automated and semi-automated resource discovery, metadata generation, and rich, full-text identification and harvest. included here are reflections on and questions about some of the probable implications and impacts of, as well as roadblocks to, machine-learning technologies applied to collection building. addressed are probable impacts leading to changing roles for libraries, librarian expertise, library standard vocabularies/schema, and the organizations that are the stewards of library standards. these include: ■ what might be the effect of these technologies on library operations, including changes in the areas and nature of expenditure of expertise required, shifts in amount of expertise required, and changes in divisions of labor (both human/human and human/ machine)? machine assistance in collection building | mitchell ■ what are the effects on libraries and end users when the coverage of finding-tool content can be greatly and inexpensively broadened and deepened? ■ how do current or traditional approaches to library- based practices and standards help foster or hinder these technologies? ■ how will best practices develop in regard to machine- assisted activities? ■ how do these technologies amplify and enable or simply prematurely dislodge librarian expertise? ■ who will own these technologies and tools? ■ how open to evolution are library metadata stan- dards and the organizations entrusted with their stewardship? ■ how will these technologies impact these standards? unfortunately, most of these questions will remain as questions unanswered. the few answers offered here must remain as tentative, contradictory, and flawed as those of most who dabble in the cottage industry of imagining library futures. still, in the effort to help map some of the new information landscape that is becoming apparent, these reflections, developed over the course of the last few years, may be small contributions toward defining and understanding what is coming. ■ licensing for automatic agents of libraries it will become increasingly important for libraries to develop licenses with commercial-resource vendors/pub- lishers that allow crawlers/classifiers and other automated programs, to be seen as agents of and for these libraries. it is important that automated agents be allowed to work with (e.g., create or enrich metadata and therefore increase end-user success in finding) both free and fee-based materials in much the same way that an expert bibliogra- pher, cataloger, or public-services librarian would when selecting, creating original metadata for, and providing access to a new commercially vended book intended to become part of a library or other well-defined collection. automated agents accessing and processing fee-based, internet-delivered information objects do so with the goal of improving the finding tools of the institution paying the fee to provide access for users to these objects (i.e., “library users”). thus, they are engaged in a bona fide, fair use of the material by and for the purchasing/subscribing institu- tion. the metadata and descriptive information these tools develop help make the materials they process more visible in collection/finding-tool contexts, a goal which should be desirable by all parties (i.e., end user, subscribing library, and owning author/publisher). ■ new medium, new organization, and an over-proliferation of electronic toll booths and borders another challenge is that internet access to library-col- lection contents and library catalog-described data, both free and fee-based, is becoming increasingly restricted as libraries, library service organizations, and publishers grope to create special aggregations, with exclusive access for their clienteles. countering this in their adherence to open access, have been, among others, services devel- oped by, for example, arxiv, the institute of museum and library services open archive, cdl escholarship, oaister, citeseer, and nsdl. differences in the two approaches may increasingly become an issue. on the one hand there is the broad, long- term community ethic favoring open access to an internet with few walls or borders, and authors enabled to publish directly via the internet through open eprint collections or dual commercial/personal-site publishing/copyrighting of their work. on the other hand there is the fairly nar- row definition of an internet information niche in which electronic/virtual services and collection access remain mapped restrictively to the sponsoring physical librar- ies/collections/institutions/publishers. libraries face a contradiction or tension between these two approaches. the latter mode is a natural effort to retain a tightly held clientele and access model that has characterized physi- cal libraries, reflecting narrowly conceived and decades- old organizational/budget/certification/user models of physical-library services and publisher controls. much of this practice is necessitated by commercial publishers (for whom libraries often have no alternative but to act as vectors), together with the lack of vision for and outdated stereotypes held of libraries by the larger organizations in which they find themselves. at the same time, much of the problem is also due to the inability of libraries to develop new cooperative organizational modes, models, and services that map better to the new medium, map better to new author and user benefits enabled by this medium, and that are better able to exploit fully and fluidly the new medium’s capabilities. the types of compartmen- talization of collections, access, and services needed for physical libraries and print, or necessitated by publisher restrictions, are increasingly an obstacle when projected onto internet access and service capabilities. thorough rethinking is needed, just as the educational and scholarly missions of the university as a whole must be thoroughly rethought in the light of internet-associated technologies and capabilities. while the information highway must be paid for, over-compartmentalization based on dated organizational and service models is yielding an over-multiplication of information technology and libraries | december toll booths and border crossings among aggregations and collections. an example has been the emphasis at many university of california campus libraries on the single campus opac rather than the pooling of resources across uc libraries for the strengthening and refinement of cdl’s melvyl union catalog. it is likely that with systemwide, multicampus shared resources, melvyl could improve in all respects vastly beyond the single campus opac. this is noted in the final report of the bibliographic services task force of the university of california libraries. overall, institutional parochialism can and has greatly lessened the value and fluidity of the internet as a medium for information provision. the booths and borders of tightly held collections make material harder to find, less visible, and less useful than would be true of more open, expansive collections and archives. as dempsey stated, libraries need to find “better ways to match supply and demand in the open network. . . . we need new services that operate at the network level, above the level of individual libraries.” for crawlers and classifiers, the booths and bor- ders that are proliferating in libraries can act disjunctively as barriers, reducing their performance. there are few answers to the challenges that over-pro- liferation of booths and borders represent. they are often practical solutions to immediate needs. still, projects that are exploring new avenues in organization and open, shar- able collections (and the standards they are based upon) should be further encouraged and supported community- wide. these include the open archives already mentioned and systems such as those ivia/data fountains work upon that to provide services for such collections in an open, inclusive, cooperative, participatory manner. while the answer will probably remain a mix of open (reflecting capabilities of media) and closed (reflecting organizational and vendor restraints) collections, it would be progress to move the balance point more toward the middle and away from so many booths and borders. ■ note on the related issue of meta-search libraries often respond to some of these open/closed/ multiple-collection aggregator and “brand” challenges and issues with meta-search services. meta-search can serve to mask the fundamental, growing problem of increasing booths and borders. meta-search, unlike the internet-borne conceptions of open service, collections, access, systems, software, and standards, does not really ask us to change our fundamental assumptions, organiza- tions, or data architectures to match the capabilities of the new information medium. it does not ask us to cooperate more fully and share at the level of collection and data; it also doesn’t encourage uniform-standards adoption and development. while meta-search is a fine answer to certain needs, sometimes it is used as a technical means to attempt to avoid these more fundamental issues. in addition, meta-search can be constraining for user search/access—i.e., it frequently disallows use of signifi- cant or unique search and metadata capabilities of each individual database to which it is applied. meta-search in libraries is becoming increasingly central, though it has many current operational flaws. among these flaws are: ■ simplification or dumbing-down of search in order to access lowest-common-denominator fields; ■ clumsy cross-walking among fields, or metadata ter- minologies that really are not equivalents; ■ difficulty in collating results/eliminating duplicates; and ■ difficulty of matching differing results ranking weightings/systems held by different bases. libraries emphasizing this approach may be increas- ingly themselves perceived as dumbed down by academ- ics, grad students, or serious researchers, who must reach beyond google, the opac, and meta-search search and display. instead of, or in addition to, meta-search, it might be wise to pursue more fully the hybrid database approach of combining heterogeneous records for multiple collec- tions (and multiple retrieval languages as needed) in one database. as computing power increases geometrically and price decreases drastically every couple of years, the challenge that the hybrid-database approach poses in regard to searching and maintenance of very large hybrid databases may soon become less of a problem. this power also implies that meta-search become more useful. ■ library standard controlled subject schema/vocabularies as the promise of automated and semi-automated meta- data generation and related tools becomes better known, it may be important for the community as a whole to urge our major subject vocabulary standards organizations, i.e., lc and oclc, to open more fully their standards and input in standard making for wider participation on the part of new communities of researchers, developers, and end users. both organizations maintain important library standard subject vocabularies/schema, lcsh/lcc, and ddc, and related large bibliographic databases and clas- sifier-training data embodying these standards. in this work, both organizations need to more actively seek out and encourage a wider variety of open innovation and development, both within and outside of the library community. this means involving more researchers, end users, and other perspectives in the effort of contribut- ing to the more rapid evolution of these standards in an attempt both to better meet end-user finding needs and machine assistance in collection building | mitchell to facilitate application of the standards through machine assistance. while oclc and lc have been generous in providing their data and standards for ivia research (others that have been generous with training data have been the cornell university library and cdl), most known work on these standards is funneled through their organizations, allies, and organizational filters. this is, of course, critical to a point for coordination; however, if overdone it may unnecessarily inhibit wider pollinations, new perspectives (e.g., a wider variety of linguists, computer scientists, and subject vocabulary/schema experts from other disciplines such as medicine and the sciences), decision making, and faster movement forward. informing the perspective here is that, while there are major costs involved in maintaining and coordinating these vocabularies/schemas, such costs are being borne directly or indirectly by the community in fees paid, mon- ies applied (often public monies through the large par- ticipating public university/land-grant libraries, among others), or labor volunteered/provided. lc is a public agency and oclc a corporate cooperative. in many ways then, libraries, through their metadata expert/cataloger community, should be seen as “owning,” as both co-author and funding agent, more of a share in these vocabularies (and other standards in library metadata) than their stew- arding organizations. a significant portion of the success of thousands of individual libraries is dependent on the successful evolution (replacement?) of these standards through the facilitation efforts and new roles adopted by these two organizations. ultimately, it must be recognized that in many ways, oclc and lc metadata schema and vocabularies (as well as conventions, styles, and customs in practical applica- tion) represent the codified wisdom, in the form of very large knowledge bases, of decades of resource descrip- tion practice on the part of information professionals in thousands of institutions. the library community is the co-author of these, and oclc and lc are their stewards. when viewing the community as owner, and when taking into account that the community needs to evolve more rapidly with its users to survive, then periodic clarifica- tion and renewal of the origin, intent, and understanding of the stewarding organizations and the standards they coordinate might help encourage more rapid, far-sighted change. libraries may or may not sink to the degree that this is realized. in this light, it should be noted that some communities, including path-breaking projects within nsdl, have made well-reasoned decisions not to use these library subject vocabulary standards (carl lagoze, pers. comm.). these are just recent examples, given that abstracting and indexing services/databases, for the jour- nal literature, have in most cases long ago chosen to use their own specialist vocabularies, often supplementing these by enabling key-word or natural-language searching of abstracts or complete full-text. among other core practical concerns here are that the library community’s standards may not be seen as useful and as widely applicable as other information communi- ties may desire. that is, if an important goal is to evolve and expand standards long associated with and emanating from the library community into becoming the standards of new, larger communities outside of libraries, then a more- guarded-than-not approach, which is slow to respond to early adaptors or innovators and slows sensible change, may not be the best path. here it should be said that there are significant ongo- ing efforts to overcome some of the challenges and better evolve lcsh/lcc. oclc’s faceted application of subject terminology (fast) may represent a step in the right direc- tion. having an entry-level vocabulary to translate end- user terminology to appropriate library subject standard vocabulary terms would be of great importance to most types of end user. oclc has also been working with the resource description network (rdn) to streamline ddc application. there just need to be more of these efforts moving at a more rapid clip. as macewan concluded in , “if lcsh does not change it will sooner or later be abandoned. . . .” the same might be said of library subject vocabulary/classification standards. however, in the worst-case scenario, assuming the existing subject standards cannot evolve more rapidly to meet new user needs in information access, collec- tion building, and metadata creation, now may even be an appropriate juncture for a large-scale rethinking and rebuilding, from the ground up. the architecture, intent, end-user audience, form, and substance of these standards would need to be rebuilt and expanded. a capability for organizationally responding more quickly to what has amounted over the last few years to far-reaching paradigm shifts would be enabled. now may be the time because, in addition to the questions of the openness/innova- tion/evolutionary adaptability of these standards, they exhibit significant, long-noted, functional flaws in terms of a non-librarian end user finding success. among others often noted are: ■ misuse/lack of understanding on the part of end users (and, rarely, poor learning materials and guidance sup- plied by librarians) due to real or perceived complex- ity, often associated with the use of subheadings and arcane terms that are far from intuitive for users). ■ typically sparse application that doesn’t fully repre- sent the number or depth of topics addressed by a work. despite the time needed to create the marc record manually, very few lcshs are applied (often three or less in the university of california’s melvyl union catalog). ■ the arcane and overly general nature of many terms that sometimes do not accord with terminology used by practitioners in the field. information technology and libraries | december ■ the lack of currency of terms describing new or recent phenomenon (see discussion of entry vocabu- lary. ■ the lack of uniformity of subject granularity in their application across multiple cataloging institutions for the same/similar works. ■ the significant amounts of expensive expert labor involved in their application. ■ their complexity often at least partially assumes some expert mediation (that may not be available, given that access is increasingly from outside the library) or long-term experience with the vocabu- lary. ■ overdone detail/complexity, some of it either not extremely useful to researchers and nonlibrarian end users or already instantly verifiable by users. ■ their arcane-ness and complexity, which limits capa- bilities for machine assistance in application and, thus thwarts a major, inexpensive means for future collection growth, increased coverage, and more use- ful collections. fortunately, and this is crucial, it turns out that much of the tonic needed for improvement may reside in the areas of inexpensively augmenting, as opposed to changing, the lcsh/lcc/ddc schema/vocabularies. for example, it is probable that most significant objects, when not digitized themselves, will be accompanied increasingly by digitized, representative cores of searchable natural-language rich text, as lc is doing with its table of contents digitization. automated and semi-automated tools for rich-text iden- tification, extraction, and end-user searching are showing applicability now (see part i). similarly, keyphrase identifi- cation and application can be accomplished automatically with a good degree of reliability; these processes play a role similar to rich text in providing useful retrieval terms and in augmenting subject searching with/without these controlled vocabularies. finally, reasonably good overall subject gisting is occurring in the creation of annotation- like constructs. all of these—rich text, keyphrases, and annotation-like constructs alike—are of great potential value in addressing controlled subject vocabulary/schema inadequacies and in complementing lcsh/lcc/ddc in end-user finding. it is also probable that use of machine means to aug- ment overarching standard subject vocabularies with complementary and much more granular/detailed spe- cialist vocabularies (both expert created and controlled as well as those that are automatically invoked) will shortly be practical and prove very useful. streamlined lcsh/ lcc/ddc could be made perhaps to function as linguistic “switching yards” with specialist vocabularies oriented to them and acting as extensions via the spine provided by the generalist vocabularies (similar to work being explored by vizine-goetz). all of this could be hinged on the syn- onomy and other term/concept relationships supplied by wordnet or other whole natural-language corpora. in such a manner, reconceived lcsh/lcc/ddc can basi- cally work as multi-vocabulary integration and translation tools in cases where the granularity of the subject becomes very fine-grained or specialized. such synonymy, lin- guistic linkages, and switching capabilities would make possible more meaningful and accurate interrelations and more fluid user movement among the vocabularies and concepts of multiple disciplines and multiple-controlled vocabularies/schema. this would also better enable the end user when employing terms actually used by practi- tioners/researchers/students in their disciplines. these and other efforts are crucial because, despite their problems, lcsh/lcc/ddc are comprehensive, overarch- ing vocabularies and schema that, though complex (as are the subject vocabularies of biosis and pubmed/medline, which successfully represent very large subject universes of their own), have done a generally useful job of repre- senting and coherently organizing finding terminology for most known worldly (and unworldly) phenomena. this, on any basis, is no easy task. these library standard vocabularies might best be seen as both essential connective tissue and as spines that could coherently thread many disciplines and interests, and many of the more specific vocabularies, together. without such a spine, interdisciplinarians, researchers/students new to an area, and generalists—whose focus requires wide knowledge often across among many disciplines (and therefore subject vocabularies)—may find themselves handicapped. each sub- and then sub-sub-specialization might develop its own mutually exclusive and contra- dictory terminology in a manner that natural-language substitutions such as keyphrase and rich-text availability can only partially fix. many end users and librarians noted the downsides of natural-language-text-only searching two decades ago while using newspaper and other full-text databases offered by dialog or brs. finally, one cannot ignore that lcsh/lcc/ddc have huge established bases of practitioners and metadata records employing them. therefore, their value is large. to summarzie, the solutions to the problems inherent in using library standard subject vocabulary/schema and other controlled metadata will involve the following: ■ openness to extensive hybridization of approaches to rethinking subject vocabularies/schema and other metadata; ■ awareness of, design for, guidance of, and incorpora- tion of new machine-assisted technologies to boost collection coverage and reduce costs of application; ■ embracing machine assistance, as appropriate, as a means of amplifying and extending expertise and application; ■ applying existent technologies for generation of key- machine assistance in collection building | mitchell phrases, description-like constructs, and rich text in order to augment controlled subject vocabularies; ■ developing a better conception of end-user metadata expectations and needs against the backdrop and expectations generated by the web, such as instant end-user access/verification; and ■ making use of specialist vocabularies that might be dovetailed well with and coordinated through stan- dard vocabularies. ■ invoked subject vocabularies—hierarchical and otherwise it is important to track recent research into automated and semi-automated means for creating (often referred to in the computer-science literature as “inducing” or extracting) hierarchical and other subject vocabularies/ontologies from natural-language corpora (see part ii). the intent of this work is to have the natural-language terms used by practitioners directly populate and structure the sub- ject-finding approach. automated induction of subject vocabularies will be useful to augment and increase the capabilities, flexibility, and interactivity of standard subject vocabularies/schema. at the very least, and this is important, they could func- tion to automatically suggest synonyms or new terminology for ongoing vocabularies/schema. and these approaches could be put to use in building entry-level vocabularies that front the vocabularies of the standards. they could also be used to aid in the semi-automated or automated repopulation/reworking of the standards, if large-scale, from-the-ground-up reworking is deemed necessary at some point. this would be done on a disci- pline-by-discipline, subject-by-subject basis. ■ resource discovery, search engines, and your library’s subject portal library collections, virtual libraries, portals, and internet- enabled catalogs of openly accessible, significant internet resources all function as “hubs” (see part i). along with other types of expert-created hubs, they have played a role in providing most large, sophisticated, commercial search engines with a significant means for modeling and determining high-quality resources and, when accurate, a considerable portion of their accuracy. though google and others do not detail how their search algorithms work, most advanced crawlers highly weight (give authority to) sites that contain large numbers of links to research and other significant resources, especially when expert created. similarly, resources from specific domains such as .edu, .org, and .gov, and institutions such as libraries, universi- ties, and scholarly societies can be identified and more highly weighted. this is another case of the community’s expertise/authority functioning as a knowledge base that, when offered as a public good (as library-created hubs often are), helps better enable directional tools for these commercial and noncommercial crawlers. there is nothing wrong with this as long as the community is aware of its contribution and as long as its efforts are recognized by these businesses. expert library-based subject portals often reciprocate usage by using commercial engines for resource discovery, though this usually represents a minor way of collecting because other expert sources are preferred. ■ enumeration of catalysts for, impacts of, and issues in machine assistance in the library community related to these research and technical developments, the library community needs to think through a great many interrelated and diverse issues and questions regarding ( ) impacts of the machine assistance we have been dis- cussing; ( ) the possible massive automation of metadata generation and resource discovery in libraries, ( ) who will “own” these technologies and ideas, and ( ) changes in expectations/roles of metadata practitioners and standards and their stewards, in the following areas: ■ when will machine learning/machine assistance yield reliable, inexpensive, and therefore massive application of metadata on an internet scale, that meets librarian, and more importantly, end-user expectations in terms of usefulness? machine assis- tance should begin to be factored into long-term planning. ■ what will be the effects of this machine amplification in changing the importance/roles/content of subject standards? that is, how and to what degree will a new means and scale of application change these standards generally, and how they’re perceived and used by end users and librarians and, therefore, be applied by the library community? how might these standards themselves change both in terms of changes in and approaches to vocabulary and schema? that is to say, how would massive, machine-assisted appli- cation in and of itself change the makeup of the vocabulary, schema, and the styles/conventions with which they are applied? ■ how might the roles of the stewards of these stan- dards change, given massive application as well as possible interest on the part of other communi- ties? can library standards penetrate and be effec- tively used by other information communities? what changes in the standards would be required to achieve this? information technology and libraries | december ■ what are the trade-offs between highly manual or craftsman/guild approaches and highly automated or more industrial approaches to applying meta- data? within which contexts, collections, resources, and budgets are these approaches to be best used, either singly or combined in various proportions, in building/expanding a collection? how does each approach best complement the other in library collec- tions? ■ to what degree will changing end-user information usage and access patterns change approaches in regard to collection design and access assumptions, the metadata standards the collections are based upon, and the stewarding organizations of the stan- dards? ■ to what degree may labor and resource savings, as well as the ability to provide for more comprehen- sive collections, as offered by this technology, dictate changes within the library community in regard to expectations for metadata quality and specificity? in which information-seeking contexts and collections and to what degree will the google-type record or minimal, streamlined dc become, if not a necessity themselves, then a pole toward which library biblio- graphic metadata evolves? ■ a question self-evident to most but not to all is: to what degree will the nature of the internet itself continue to change our approach to supplying meta- data? again, researchers in academic departments no longer need walk across campus to the library by virtue of having many bibliographic details of an object present in a metadata record. increasingly, they can go to the object on the internet and instantly verify the detail for themselves. should libraries de- emphasize data elements/fields that are dependably and quickly end-user verifiable in favor of expend- ing more expertise, time, and resources in gisting/ describing the subject, intent, and perhaps even esti- mated quality or significance of the work? ■ in which specific ways will labor be saved and machines be capable of assisting in resource discov- ery and metadata generation? that is, what level of automation/semi-automation is acceptable to the community and reliably deployable in production over horizons of one to five years? what level of qual- ity/depth will users accept in metadata designed to occupy the continuum existing between the marc record and the google “record” (this being a large and significant service area; see part i)? how will this technology change old and enable new roles, tasks, and production routines for library subject experts and other staff? how will libraries ramp up and tran- sition into this? ■ will the substantial potential economic advantages of automated or semi-automated generation of library standard metadata such as lcsh/lcc/ddc vocab- ularies/schema drive a rethinking toward greater uniformity/simplicity/streamlining of these stan- dards and conventions in their application, explicitly with machine application in mind? for example, per- haps only a subset of a whole vocabulary will be used and those that are used will become less detailed and less rich for experts but also—for most end users— less complex and arcane, and more intuitive. ■ in some ways, the existence of dc is a recognition that this kind of rethinking and streamlining of library description standards, in the interest of repre- senting and providing access to a much larger scale of communities and resources, is already well under way. what are the obstacles to greater usage of dc? ■ what should the balance be in streamlining metadata for automated application, in relation to its cur- rent complexity/depth while augmenting with rich text? from another perspective, what is the balance when considering the oversimplification and loss of descriptive power when using machine methods as compared with that otherwise achievable through use of subject expertise? how will libraries deter- mine best balances of expert and machine in regard to different tasks? how will this be quantified and determined through examination of user retrieval success/satisfaction—with this, in turn, factored against the backdrop of metadata creation costs, full- text data harvesting and retrieval, and the need for collections with much greater reach? ■ as accurate means of metadata and rich-text gen- eration for/from text objects improve, machine assis- tance will allow a shifting of expertise to provide better collection coverage and expression of subject- domain expertise (e.g., in abstracts). how will this new capability for breadth and depth be defined and used in library collections? for example, will new visual, multimedia, and data objects—which the web has made possible on a mass basis and which librar- ies generally do not cover well—become a major goal in repurposing expertise since these do not eas- ily lend themselves to machine processing (karen calhoun, pers. comm.)? ■ might streamlining and the usage of multiple depths/types of metadata application first require the acceptance within the community of the concept of the multitiered collection/database that supports multiple levels and types of heterogeneous resources representing differing levels of importance to users? or, can this need be met through more fully evolved meta-search approaches? ■ helping to structure this metadata heterogeneity might be the sliding-scale application of varying levels of metadata-generation labor expenditures and amounts/type of metadata, with the lower- machine assistance in collection building | mitchell and middle-value resources receiving application of streamlined standard vocabularies/schema and rich text, automatically or semi-automatically, at low cost. high-value resources would continue to receive expert-applied, expensively created, com- plex, and high-quality metadata as well as rich text. libraries already make such distinctions in qual- ity/significance to some degree through purchasing (e.g., departmental collecting profiles/weightings by subject and object type and cost) and order-of- cataloging priority decisions, as well as by student/ faculty input on specific items. more specifically, we would need to discuss and develop criteria in deter- mining the core or peripheral value of a resource for its subjects and user communities and then, based on the judgments derived, appropriately apportion amount and type of metadata and expert labor or machine assistance, on a sliding scale. again, while it should be noted that the library community has gen- erally avoided rendering judgments on the possible use/relevance of a resource to a subject community, libraries nevertheless do routinely make general calls that effectively function this way to some degree. in making this judgement, it would be critical to involve resource users. reviewer-researcher, library user, and librarian evaluations for purchases as well as find- ing tool/collection-usage statistics for the specific subject or author and item all could be woven into the means by which the core weighting of a resource could be assigned and be refined over time via usage. developing this value is important from a library standpoint. it is a key that may help unlock solu- tions for some of the community’s bigger challenges, including those revolving around the best marriage of machine assistance with librarian expertise. how do libraries go about making these sliding-scale evaluations with some uniformity, among different collection types and interests, with an eye toward tasking expert and machine? ■ can some of the general end-user search deficiencies commonly acknowledged for lcsh/lcc/ddc be rectified to some extent by automatically/semi-auto- matically providing rich full-text accompaniment for each record/resource, either in the form of “selected” excerpts verbatim or as processed into significant key- phrases representing this text? how could the pres- ence of this rich text not so much change as augment these standards? for example, rich full-text might be relied upon to contain detail that obviated the need to use certain lcsh subdivisions or other types of marc metadata. could inadequacies/inaccuracies in expert-applied and machine-applied metadata be partially countered, for end-user retrieval purposes, through the presence of rich full-text? rich text, as well as keyphrases/terms and descriptions that serve the same purpose in this context, can now be reliably generated in many cases automatically. what would be the right mix of subject-vocabulary standard meta- data and accompanying, selected natural-language text for best end-user success? how might rich-text extraction and searching improve upon searching of whole-object full-text? how much rich text is needed and how distilled should it be? large, whole-object full-text searching can often be a searcher’s quag- mire, clouding results rankings and weightings. ■ could a new scale of application and interest on the part of new communities be better catalyzed through the incentive offered by opening up the lcsh/lcc/ ddc subject vocabularies/schema on an open-stan- dards/open-source, free-software model? ■ if development of these technologies is constrained with regard to action/inaction on the part of the community and its stewards, will the standards be replaced—or become obsolete—for major existing or prospective sectors of users? if so, what does this mean for the library community? ■ by and for whom is such standard subject vocab- ulary/schema application technology developed within the community? classifiers are actually trained through great amounts of what, in many cases, is really community-created knowledge in order to apply community-developed schema/vocabularies. smart crawlers and extractors similarly use (have “learned”) collectively created information patterns, derived from open-knowledge bases of various sorts. who should own these tools/models and how open/ closed should the programming code/ideas be, con- sidering they could not be built without using the collective wisdom embodied in these knowledge bases? these tools exploit decades of labor by thou- sands of institutions, whose assumption has gener- ally been that the knowledge base and, by extension, the tools that are built on and benefit from it, are and should remain directly or indirectly, public goods. ■ for whom is machine learning/assistance in collec- tion building patented? the ideas, training corpora, algorithms, and data models discussed need to be observed and protected for the public domain to encourage their widespread and inexpensive avail- ability, as well as their evolution. the u.s. patent and trademark office is now more commonly supporting the patenting of whole, generic processes that have heretofore had one or both feet in the commons, as compared with solely granting patent rights in more discrete areas of original invention. it would be unfortunate to find one day that machine assistance in collection building had been patented. this is especially an issue, given that there is little machine learning of interest to libraries that does not mine, apply, and extend the stored wisdom and knowledge that the community has built for decades. information technology and libraries | december ■ summary of part iii it is important to think through and anticipate a great number of issues and concerns—including those of open models and open development—regarding machine-assis- tance tools (e.g., classifiers, extractors, and related algo- rithms/models) that generate library standard metadata, and identify and extract useful natural-language data. it is important because these tools could become central activi- ties in libraries over the next one to five years. reflection here is especially appropriate, given the degree that these tools are trained on exemplars from library collections and come to distill and embody models of library metadata, standards, and expertise that represent the knowledge created over decades through the effort of a whole com- munity. it is important to think through what machine- assistance technologies in collection building imply for the future role of the librarian’s expertise. specifically, libraries need to reconceptualize machine-assistance software not as fully automated “ai” but rather, as enabling expert driven, strongly interactive, “servo-mechanisms” that semi-automate some work to increase the reach, quality, and user-finding success within library collections. while it will probably start out with ten or fifteen minutes of expert time saved per record by such tools, this is a lot of time saved when aggregated across the entire community and will only increase. and the community needs to think through what this implies for the evolution of library- standard metadata, given that machine assistance will increasingly allow for massive and economic application, if a convergence of machine capabilities and machine- friendly metadata standards is architected. this large-scale amplification of usage will quite likely involve changing the value/roles of these standards for the community, as well as for the larger communities that may come to use them at the cost of simplification, streamlin- ing, and a greater reliance on end users to verify some of their own metadata details (often interacting directly with the digital resource). the tools also imply a restructuring of expertise and its application in metadata creation in libraries to reflect a division of labor, with semi-automated machine description processes spent on the mass of useful but mid- to lower-value materials; with and expert time being spent on high-value resources; and with both types of records residing in the same multitiered, heterogeneous collection. finally, needing examination will be the roles of the stewardship organizations in: ■ shepherding the community’s metadata standards during a period of great change; ■ openly evolving the application of metadata stan- dards within the context of machine assignment for the greatest possible good; ■ rapidly evolving the application of metadata stan- dards to retain guidance of and to keep pace with open and proprietary developments in these areas; ■ distilling the metadata knowledge base and wisdom created by the community as this is transformed into the programmatic knowledge (rule bases and mod- els) used by new tools. this knowledge base is a priceless asset for the library community in sustaining service roles in an age of the large-scale advent of commercial-information access, delivery, and ownership. ■ conclusion this article discusses work over the last several years in machine-learning software and services relevant to collection building in libraries. a number of promising avenues for exploration and research are detailed. deeper understanding of and more direct involvement in areas of machine learning are urged for libraries in order to reflect advances in the computer sciences and other disciplines as well as to meet changing end-user needs among infor- mation seekers. ■ acknowledgements the author would like to thank the u.s. institute of museum and library services; the library of the university of california at riverside; the national science foundation’s national science digital library; the fund for the improvement of post-secondary education of the u.s department of education; the librarians association of the university of california; and the computing and communications group of the university of california at riverside for current or past funding support. the author would also like to thank the library of congress; cornell university library; oclc; and the california digital library for providing training data and other assistance for the research. thanks to karen calhoun (cornell university library) and two anonymous readers for some excellent comments and suggestions. finally, the author would like to commend ivia lead programmer johannes ruscheinski, primary author of the data fountains and ivia code bases, for his excellent work over the years, as well as gordon paynter, walt howard, jason scheirer, keith humphries, anthony moralez, paul vander griend, artur kedzierski, margaret mooney, john saylor, laura bartolo, carlos rodriguez, jan herd, carolyn larson, diane hillmann, and ruth jackson for their invaluable contributions to the machine assistance in collection building | mitchell projects. the views expressed here are solely those of the author and not intended to represent those of the library of the university of california, riverside, our funding agencies, or cooperators. ■ references and notes . s. mitchell et al., “ivia: open source virtual library software,” d-lib magazine (january ). http://www.dlib .org/dlib/january /mitchell/ mitchell.html (accessed oct. , ); g. paynter, “developing practical automatic meta- data assignment and evaluation tools for internet resources,” in proceedings of the th acm/ieee joint conference on digital libraries (denver: acm pr., ), – (winner of the jcdl vannevar bush best paper award), http://ivia.ucr.edu/ projects/publications/paynter- -jcdl-metadata-assign- ment.pdf, (accessed oct. , ); s. mitchell, “collaboration enabling internet resource collection-building software and technologies,” library trends , no. (may ): – ; j. mason et al., “infomine: promising directions in virtual library development,” first monday ( ), http://www.first monday.dk/issues/issue _ /mason/ (accessed oct. , ). . s. mitchell, “infomine: the first three years of a vir- tual library for the biological, agricultural, and medical sci- ences,” in proceedings of the contributed papers session, biological sciences division, special libraries association annual conference (seattle: special libraries assocation, ). . mitchell, “collaboration enabling internet resource col- lection-building software and technologies.” . j. phipps et al., “orchestrating metadata enhancement services: introducing lenny,” in proceedings of dc- : inter- national conference on dublin core and metadata applications (madrid, spain: universidad carlos iii de madrid, ), http://arxiv.org/pdf/cs.dl/ , (accessed oct. , ). . mason et al., “infomine: promising directions in vir- tual library development.” . ibid. . s. chakrabarti, mining the web: discovering knowledge from hypertext (san francisco: morgan kauffman, ); s. chakrabarti et al., accelerated focused crawling through online relevance feed- back, http://www .org/cdrom/ refereed/ / (accessed oct. , ); s. chakrabarti, the structure of broad topics on the web, http://www .org/cdrom/refereed/ /index.html (accessed oct. , ); s. chakrabarti, integrating the document object model with hyperlinks for enhanced topic distillation and information extraction, http://www .org/cdrom/papers/ (accessed oct. , ). . chakrabarti et al., accelerated focused crawling; f. menc- zer, “mapping the semantics of web text and links” iee internet computing, , no. (may/june ): – ; f. menczer, g. pant, and p. srinivasan, “topical web crawlers: evaluating adaptive algorithms” transactions on internet technology, , no ( ): –; f. menczer, “correlated topologies in citation networks and the web” european physical journal b, no. (march ): – . . s. mitchell, “data fountains survey,” , http:// datafountains.ucr.edu/ datafountainssurvey.doc, (accessed oct. , ). . a. culotta and a. mccallum, “confidence estimation for information extraction,” in proceedings of human language technology conference and north american chapter of the asso- ciation for computational linguistics (boston: association for computational linguistics, ), http://www.cs.umass.edu/ ~mccallum/papers/crfcp-hlt .pdf, (accessed oct. , ); f. peng and a. mccallum, “accurate information extraction from research papers using conditional random fields,” in pro- ceedings of the human language technology conference and north american chapter of the association for computational linguistics ( ). http://ciir.cs.umass.edu/pubfiles/ir- .pdf, (accessed oct. , ); c. sutton and a. mccallum, “an introduction to conditional random fields for relational learning,” in introduction to statistical relational learning, lise getoor and ben taskar, eds. (cambridge, mass.: mit pr., ). http://www .cs.umass.edu/~mccallum/papers/crf-tutorial.pdf, (accessed oct. , ). . a. mccallum and d. jensen, “a note on the unification of information extraction and data mining using conditional- probability, relational models,” in proceedings of the ijcai workshop on learning statistical models from relational data, aca- pulco, mexico: ijcai, http://www.cs.umass.edu/~mccallum/ papers/iedatamining-ijcaiws .pdf, (accessed oct. , ); u. nahm and r. mooney, “a mutually beneficial integration of data mining and information extraction,” in proceedings of the american association for artificial intelligence/innovative applica- tions of artificial intelligence (austin, texas: american asso- ciation for artificial intelligence, ). http://www.cs.utexas .edu/users/ ml/papers/discotex-aaai- .pdf, (accessed oct. , ); r. raina et al., “classification with hybrid generative/ discriminative models,” in proceedings of neural information pro- cessing systems ( ). http://www.cs.umass.edu/~mccallum/ papers/hybrid-nips .pdf, (accessed oct. , ); g. bouchard and b. triggs, “the trade-off between generative and discrimi- native classifiers,” compstat . (prague: springer, ) http://lear.inrialpes.fr/pubs/ /bt /bouchard-comp stat .pdf, (accessed oct. , ). . mccallum and jensen, “a note on the unification of information extraction.” . n. eiron and k. mccurley, “untangling compound docu- ments on the web,” in conference on hypertext (nottingham, uk: acm conference on hypertext and hypermedia, ), http://citeseer .ist.psu.edu/eiron untangling.html, (accessed oct. , ). http://www.almaden.ibm.com/cs/people/mccurley/pdfs/ pdf.pdf, (accessed oct. , ); p. dimitriev et al., “as we may perceive: inferring logical documents from hypertext,” presented at ht , th acm conference on hypertext and hypermedia (salzburg: acm, ); k. tajima, “finding context paths for web pages,” in proceedings of acm hypertext (darm- stad, germany: acm, ), http://www.jaist.ac.jp/~tajima/ information technology and libraries | december papers/ht www.pdf, (accessed oct. , ); k. tajima et al., “discovery and retrieval of logical information units in web,” in proceedings of the workshop of organizing web space (in conjunc- tion with acm conference on digital libraries) (berkeley, calif.: acm, ), – , http://www.jaist.ac.jp/~tajima/ papers/ wows www.pdf, (accessed oct. , ); e. de lara et al., “a characterization of compound documents on the web,” tr - , university of toronto ( ), http://www.cs.toronto .edu/~delara/papers/compdoc.pdf, (accessed oct. , ), http://www.cs.toronto.edu/~delara/ papers/compdoc_html/, (accessed oct. , ); l. xiaoli et al., “web search based on micro information units,” (honolulu, hawaii: eleventh inter- national world wide web conference, ), http://www .org/cdrom/poster/ .pdf, (accessed oct. , ); w. lee et al., retrieval and organizing web pages by information unit, http://www .org/cdrom/papers/ /, (accessed oct. , ). . tajima et al., “discovery and retrieval of logical informa- tion units in web”; xiaoli et al., “web search based on micro information units”; lee et al., retrieval and organizing web pages. . r. mihalcea, “graph-based ranking algorithms for sen- tence extraction, applied to text summarization,” in proceedings of the nd annual meeting of the association for computational linguistics, companion volume (barcelona, spain: associa- tion for computational linguistics, ), http://www.cs.unt .edu/~rada/papers/mihalcea.acl .pdf, (accessed oct. , ); r. mihalcea and p. tarau, “textrank: bringing order into texts,” in proceedings of the conference on empirical methods in natural language processing (barcelona, spain: empirical meth- ods in natural language processing, ), http://www.cs.unt .edu/~rada/papers/mihalcea.emnlp .pdf, (accessed oct. , ); r. mihalcea, p. tarau, and e. figa, “pagerank on semantic networks, with application to word sense disambiguation,” in proceedings of the th international conference on computational linguistics (geneva, switzerland: coling ). http://www .cs.unt.edu/~rada/papers/ mihalcea.coling .pdf, (accessed oct. , ); y. matsuo et al., “keyworld: extracting keywords in a document as a small world,” in proceedings of discovery sci- ence (berlin, new york: springer, ), – (lecture notes in computer science, v. ), http://www.miv.t.u-tokyo.ac.jp/ papers/matsuods .pdf, (accessed oct. , ); y. matsuo and m. ishizuka, “keyword extraction from a single document using word co-occurrence statistical information,” interna- tional journal on artificial intelligence tools , no. ( ): – , http://www.miv.t.u-tokyo.ac.jp/papers/matsuoijait .pdf, (accessed oct. , ); xiaoli et al., “web search based on micro information units”; lee et al., retrieval and organizing web pages; g. forman and ira cohen, “learning from little: com- parison of classifiers given little training,” tech report: hpl- - r (palo alto, calif.: hewlett-packard research labs., ), http://www.hpl.hp.com/techreports/ /hpl - - r .pdf, (accessed oct. , ). . g. mann et al., “bibliometric impact measures leveraging topic analysis,” (in press), in proceedings of the joint conference on digital libraries ( ). http://www.cs.umass.edu/~mccallum/ papers/impact-jcdl s.pdf, (accessed oct. , ). . r. bouckaert and e. frank, “evaluating the replicability of significance tests for comparing learning algorithms,” in proceedings of the pacific-asia conference on knowledge discovery and data mining. (berlin, new york: springer-verlag, ), – (lecture notes in computer science, v. ), http://www .cs.waikato.ac.nz/~ml/publications/ /bouckaert-frank.pdf, (accessed oct. , ); r. bouckaert, “estimating replicabil- ity of classifier learning experiments,” in proceedings of the international conference on machine learning ( ), http://www. cs.waikato.ac.nz/~ml/publications/ /bouckaert-estimat- ing.pdf, (accessed oct. , ); r. caruana and a. niculescu- mizil, “data mining in metric space: an empirical analysis of supervised learning performance criteria,” in kdd- : proceedings of the tenth acm sigkdd international conference on knowledge discovery and data mining (new york: acm press, ), http://perfs.rocai .revised.rev .ps, (accessed oct. , ). . j. zhang et al., “modified logistic regression: an approx- imation to svm and its application in large-scale text cat- egorization,” in proceedings: twentieth international conference on machine learning (menlo park calif.: aaai press, ), – , http://www.informedia.cs.cmu.edu/documents/icml zhang .pdf, (accessed oct. , ); y-c. chang, “boosting svm classifiers with logistic regression,” technical report. (tai- pei: institute of statistical science, academia sinica, ), http://www.stat.sinica.edu.tw/library/c_tec_rep/ - .pdf, (accessed oct. , ); t. zhang and f. oles, “text categori- zation based on regularized linear classification methods,” information retrieval , no. ( ): – , http://www.research .ibm.com/people/t/tzhang/pubs.html, (accessed oct. , ); t. joachims, “svmlight,” (including svmmulticlass, svmstruct, svmhmm) (software, ), http://svmlight.joachims.org/, (accessed oct. , ); c. chang and c-j. lin, “libsvm,” (software, ), http://www.csie.ntu.edu.tw/~cjlin/libsvm/, (accessed oct. , ); c-w hsu and c-j lin, “bsvm,” (soft- ware, ), http://www.csie.ntu.edu.tw/~cjlin/bsvm/index .html, (accessed oct. , ); t. finley and t. joachims, “supervised clustering with support vector machines,” in proceedings of the international conference on machine learning (new york: acm press, ), http://www.cs.cornell.edu/ people/tj/publications/finley_joachims_ a.pdf, (accessed oct. , ); i. tsochantaridis et al., “support vector machine learning for interdependent and structured output spaces,” in proceedings of the international conference on machine learning (new york: acm press, ), http://www.cs.cornell.edu/ people/tj/publications/tsochantaridis_etal_ a.pdf, (accessed oct. , ) ; s. godbole and s. sarawagi, “discriminative methods for multi-labeled classification,” in proceedings of the pacific-asia conferences on knowledge discovery and data min- ing ( ), http://www.it.iitb.ac.in/~shantanu/work/pakdd .pdf, (accessed oct. , ); l. cai and t. hofmann, “hierarchi- cal document categorization with support vector machines,” in proceedings of the acm th conference on information and knowl- machine assistance in collection building | mitchell edge management ( ), http://www.cs.brown.edu/people/ th/publications.html, (accessed oct. , ); t. hofmann et al., “learning with taxonomies: classifying documents and words,” in proceedings of the workshop on syntax, seman- tics, and statistics, neural information processing ( ), http:// www.cs.brown.edu/people/ th/publications.html, (accessed oct. , ); a. tveit, “empirical comparison of accuracy and performance for the mipsvm classifier with existing classifiers,” technical report, division of intelligent systems, department of computer and information science, norwegian university of science and technology. (trondheim, norway, ), http://www.idi.ntnu.no/~amundt/publications/ / mipsvmclassificationcomparison.pdf, (accessed oct. , ); c-w hsu and c-j lin, “a comparison of methods for multi- class support vector machines,” ieee transactions on neu- ral networks , no. ( ): – , http://www.csie.ntu .edu.tw/~cjlin/papers/multisvm.pdf, (accessed oct. , ). . p. komarek, “logistic regression for data mining and high-dimensional classification” (ph.d. thesis, carnegie mellon university, ), ; p. komarek and a. moore, “fast robust logistic regression for large sparse data sets with binary outputs,” proceedings of the ninth international workshop on artificial intelligence and statistics. january – , , hyatt hotel, key west, florida, ed. by christopher m. bishop and brendan j. frey. http://research.microsoft.com/ conferences/aistats /proceedings/ .pdf (accessed nov. , ); a. popescul et al., “towards structural logistic regression: combining relational and statistical learning,” in mrdm : workshop on multi-relational data mining, http://www-ai.ijs.si/sasodzeroski/mrdm /proceed ings/popesul.pdf (accessed nov. , ); j. zhang and y. yang, “probabilistic score estimation with piecewise logistic regression,” in proceedings: twenty-first international conference on machine learning (menlo park, calif.: aaai press, ), http://www- .cs.cmu.edu/~jianzhan/papers/icml zhang .pdf, (accessed oct. , ); zhang et al., “modified logistic regression”; zhang and oles, “text categorization”; multi-class lr is discussed in zhang et al., , and chang, (reference ). . some recent work on nb can be seen in j. rennie, “tack- ling the poor assumptions of naive bayes text classifiers,” in t. fawcett and n. mishra, eds., proceedings of the th interna- tional conference on machine learning (washington, d.c.: aaai pr., ), – , http://haystack.lcs.mit.edu/papers/rennie .icml .pdf, (accessed oct. , ); k. schneider, “tech- niques for improving the performance of naive bayes for text classification,” in computational linguistics and intelli- gent text processing: sixth international conference, cicling , mexico city, mexico, february – , : proceedings (new york: springer, ). (lecture notes in computer science, ). – , http://www.phil.uni-passau.de/linguistik/ schneider/pub/cicling .html, (accessed oct. , ); e. frank et al., “locally weighted naive bayes,” in proceedings of the th conference in uncertainty in artificial intelligence (acapulco: morgan kaufmann, ), – , http://www .cs.waikato.ac.nz/~eibe/pubs/uai_ .ps.gz, (accessed oct. , ); g. webb et al., “not so naive bayes: aggregating one-dependence estimators,” machine learning , no. (jan. ): – , http://www.csse.monash.edu.au/~webb/files/ webbboughtonwang .pdf, (accessed oct. , ); e. keogh and m. pazzani, “learning the structure of augmented bayes- ian classifiers,” international journal on artificial intelligence tools , no. ( ): – , http://www.ics.uci.edu/~pazzani/ publications/tools.pdf (accessed oct. , ). . mccallum and jensen, “a note on the unification of information extraction and data mining”; joachims, “svm- light”; y. altun et al., “hidden markov support vector machines,” in proceedings of the th international conference on machine learning (menlo park, calif.: aaai press, ), http://www.cs.brown.edu/people/th/publications.html (accessed oct. , ); a. ganapathiraju et al., “hybrid svm/hmm architectures for speech recognition,” in advances in neural information processing systems : proceed- ings of the conference (cambridge, mass.: mit press, ), http://www.nist.gov/speech/publications/tw /pdf/cp .pdf (accessed oct. , ); d. freitag and a. mccallum, “information extraction with hmm structures learned by stochastic optimization,” in proceedings of the th conference on artificial intelligence (austin, tx.: aaai press, ) http:// www.cs.umass.edu/~mccallum/papers/iehill-aaai s .ps (accessed oct. , ); s. basu et al., “a probabilistic framework for semi-supervised clustering,” in proceedings of the th acm sigkdd international conference on knowl- edge discovery and data mining (seattle, wash.: ), – , http://www.cs.utexas.edu/users/ml/papers/semi-kdd - .pdf, (accessed oct. , ). . t. liu et al., “efficient exact knn and nonparametric classification in high dimensions,” in advances in neural information processing systems : proceedings of the con- ference (cambridge, mass.: mit press, ). http://www .autonlab.org/autonweb/showpaper.jsp?id=liu-knn, (accessed oct. , ); g. guo et al., “knn model-based approach in classification,” in lecture notes in computer science, vol. (heidelberg: springer berlin, ), – , http://www .icons.rodan.pl/publications/% bguo % d.pdf (accessed oct. , ) . bouckaert and frank, “evaluating the replicability of significance tests”; bouckaert, “estimating replicability of clas- sifier learning experiments”; caruana and niculescu-mizil, “data mining in metric space”; r. caruana and t. joachims, “perf (data mining evaluation software),” in proceedings of the conference on knowledge discovery and data mining ( ). http://kodiak.cs.cornell.edu/kddcup/software.html (accessed oct. , ); paynter, “developing practical automatic meta- data.” . raina et al., “classification with hybrid generative/ discriminative models”; bouchard and triggs, “the trade-off between generative and discriminative classifiers.” . ibid; zhang et al., “modified logistic regression”; chang, “boosting svm classifiers with logistic regression”; joachims, information technology and libraries | december “svmlight”; l. shih et al., “not too hot, not too cold: the bundled svm is just right!” in proceedings of the icml- workshop on text learning ( ). http://people.csail.mit.edu/ u/j/jrennie/public_html/papers/icml -bundled.pdf (accessed oct. , ); f. fukumoto and y. suzuki, “manipulating large corpora for text classification,” in proceedings of the conference on empirical methods in natural-language processing (philadel- phia: association for computational linguistics, ), – , http://acl.ldc.upenn.edu/w/w /w - .pdf (accessed oct. , ); altun et al., “hidden markov support vector machines; ganapathiraju et al., “hybrid svm/hmm architec- tures”; liu et al., “efficient exact k-nn”; a. ng and m. jordan, “on discriminative versus generative classifiers: a com- parison of logistic regression and naive bayes,” in advances in neural information processing systems : proceedings of the conference (cambridge, mass.: mit press, ), http:// www.robotics.stanford.edu/~ang/ papers/nips -discriminati- vegenerative.ps (accessed oct. , ); k. nigam et al., “text classification from labeled and unlabeled documents using em,” machine learning , nos. / ( ): – , http://www .kamalnigam.com/papers/emcat-mlj .pdf (accessed oct. , ). . g. valentini and f. masulli, “ensembles of learning machines,” in neural nets wirn vietri- , series lecture notes in computer sciences, m. marinaro and r. tagliaferri, eds. (heidelberg: springer-verlag, ), http://www.disi.unige.it/ person/masullif/papers/masulli-wirn .pdf (accessed oct. , ). . ibid.; r. caruana et al., “ensemble selection from librar- ies of models” in proceedings: twenty-first international conference on machine learning (menlo park, calif.: aaai press, ). http://www.cs.cornell.edu/~alexn/shotgun.icml .revised. rev .pdf (accessed oct. , ); g. tsoumakas, “effective vot- ing of heterogeneous classifiers,” in machine learning ecml : th european conference on machine learning, pisa, italy, september – , : proceedings. (berlin, new york: springer, ), http://users.auth.gr/~greg/publications/tsoumakas -ecml .pdf (accessed oct. , ); j. fürnkranz, “on the use of fast sub-sampling estimates for algorithm recommen- dation,” technical report tr- - (wien: Österreichisches forschungsinstitut für artificial intelligence, ), http://www .ofai.at/cgi-bin/get-tr?paper=oefai-tr- - .pdf (accessed oct. , ); a. seewald, . “meta-learning for stacked classifi- cation,” (extended version) in proceedings of the nd international workshop on integration and collaboration aspects of data mining, decision support, and meta-learning (university of helsinki, department of computer science, report b- - , ), http:// www.ofai.at/cgi-bin/get-tr?download= &paper=oefai-tr- - .pdf (accessed oct. , ); a. seewald and j. fürnkranz, “an evaluation of grading classifiers,” in advances in intelli- gent data analysis: proceedings of the th international symposium (lisbon, portugal: springer-verlag, ), http://www.ofai.at/ cgi-bin/get-tr?paper=oefai-tr- - .pdf (accessed oct. , ); p. bennett et al., “the combination of text classifiers using reliability indicators,” technical report. microsoft and information retrieval , no. ( ): – , http://research .microsoft.com/~horvitz/tclass_combine.pdf (accessed oct. , ); y. kim et al., “optimal ensemble construction via meta- evolutionary ensembles,” expert systems with applications , no. (in press ), http://www.informatics.indiana.edu/fil/ papers/mee-eswa.pdf (accessed oct. , ). . s. godbole, “document classification as an internet ser- vice: choosing the best classifier” (masters thesis, iit bombay, ). http://www.it.iitb.ac.in/~shantanu/work/mtpsg.pdf (accessed oct. , ). . k. liu and h. kargupta, “distributed data mining bibli- ography: release . ,” (baltimore: university of maryland, com- puter science department, ), http://www.csee.umbc.edu/ ~hillol/ddmbib/ (accessed oct. , ); a. prodromidis and p. chan, “meta-learning in distributed data mining systems: issues and approaches,” in advances of distributed data mining, hillol kargupta and philip chan, eds. (menlo park, calif. : aaai/ mit press, ). http://www .cs.columbia.edu/~andreas/ publications/ddmbook.ps.gz (accessed oct. , ); g. tsoumakas and i. vlahavas, “distributed data mining of large classifier ensembles,” in methods and applications of artificial intelligence: second hellenic conference on ai, setn , thes- saloniki, greece, april – , : proceedings, (berlin, new york: springer, ), – , http://users.auth.gr/~greg/pub- lications/ddmlce.pdf (accessed oct. , ); r. khoussainov et al., “grid-enabled weka: a toolkit for machine learn- ing on the grid,” ercim news no. , (oct. ), http:// www.ercim.org/publication/ercim_news/enw /khussainov .html (accessed oct. , ). . s. godbole et al., “document classification through inter- active supervision of document and term labels,” in knowledge discovery in databases: pkdd : th european conference on principles and practice of knowledge discovery in databases, pisa, italy, september – , : proceedings (berlin; new york: springer, ), http://www.it.iitb.ac.in/~shantanu/work/ pkdd .pdf (accessed oct. , ).; h. yu et al., “pebl: posi- tive example based learning for web page classification using svm,” in kdd- : proceedings of the eighth acm sigkdd international conference on knowledge discovery in data mining (new york: acm pr., ), – , http://eagle.cs.uiuc.edu/ pubs/ /pebl-kdd .pdf (accessed oct. , ); t. krist- jannson et al., “interactive information extraction with con- strained conditional random fields,” in proceedings: nineteenth national conference on artificial intelligence (aai- ) (menlo park, calif.: aaai press; cambridge, mass.: mit press, ), http://www.cs.umass.edu/~mccallum/papers/addrie-aaai . pdf (accessed oct. , ); v. tablan et al., “ollie: on-line learning for information extraction,” in proceedings of the hlt-naacl workshop on software engineering and architecture of language technology systems: edmonton, canada: . (new york: acm, ), http://gate.ac.uk/sale/hlt /ollie-sealts.pdf (accessed oct. , ). . godbole et al., “document classification.” . ibid.; tablan et al., “ollie: on-line learning for infor- mation extraction.” machine assistance in collection building | mitchell . g. mann et al., “bibliometric impact measures,” (in press). . bouckaert and frank, “evaluating the replicability of significance tests”; caruana and niculescu-mizil, “data mining in metric space”; caruana and joachims, “perf (data mining evaluation software).” . mann et al., “bibliometric impact measures”; matsuo et al., “keyworld”; matsuo and ishizuka, “keyword extraction from a single document”; lee et al., retrieval and organizing web pages; tajima et al., “discovery and retrieval of logical information.” (see also the sections on hybrid, unified models, and document scale learning and classification, above.) . menczer, “mapping the semantics of web text and links.” . p. srinivasan et al., “a general evaluation framework for topical crawlers,” information retrieval , no. ( ): – , http://www.informatics.indiana.edu/fil/papers/ crawl_framework.pdf (accessed oct. , ); a. maguit- man et al., “algorithmic computation and approximation of semantic similarity,” (in press, ). to appear in world wide web journal. http://www.informatics.indiana.edu/fil/papers/ semsim_extended.pdf (accessed oct. , ). . arxiv. cornell university library, http://arxiv.org/ (accessed oct. , ); citeseer.ist (formerly researchindex), http://citeseer.ist.psu.edu/ (accessed oct. , ); escholarship repository, california digital library, http://repositories.cdlib .org/escholarship/, (accessed oct. , ); national science foundation, national science digital library, http://nsdl.org/ (accessed oct. , ); oaister. digital library production ser- vice (university of michigan), http://oaister.umdl.umich.edu/ o/oaister/ (accessed oct. , ); u.s. institute of museum and library services. digital collections and content, http://imlsdcc .grainger.uiuc.edu/ (accessed oct. , ). . k. calhoun, “the changing nature of the catalog and its integration into other discovery tools,” (report to the library of congress, mar. , ), http://www.loc.gov/catdir/calhoun -report-final.pdf (accessed oct. , ); mitchell, “collabora- tion enabling internet resource collection-building software and technologies”; w. wulf, “higher education alert: the railroad is coming,” in educause, publications from the forum for the future of higher education ( ), http://www.educause. edu/ir/library/pdf/ffpiu .pdf (accessed oct. , ). . university of california libraries, “rethinking how we provide bibliographic services at the university of cali- fornia,” final report of the bibliographic services task force of the university of california libraries, , http://libraries .universityofcalifor nia.edu/sopag/bstf/final.pdf (accessed oct. , ). . l. dempsey, “libraries and the long tail: some thoughts about libraries in a network age,” d-lib magazine , no. ( ), http://www.dlib.org/dlib/april /dempsey/ dempsey.html (accessed oct. , ). . mason, j. et al., “infomine: promising directions in virtual library development,” first monday , no. (june , ), http://www.firstmonday.dk/issues/issue _ /mason/ (accessed oct. , ). . e. o’neill and l. m. chan, “fast: faceted application of subject terminology,” in proceedings of the world information con- gress, ifla general conference and council (berlin: ifla, ). http://www.ifla.org/iv/ifla /papers/ e-oneill_mai- chan.pdf (accessed oct. , ); see also: oclc – , “fast: faceted application of subject terminology,” http:// www.oclc.org/research/projects/fast/default.htm) (accessed oct. , ). . m. bates, , “improving user access to library cata- log and portal information,” task force recommendation . , final report (washington, d.c.:library of congress, ), , http:// www.loc.gov/catdir/bibcontrol/ . batesreport - .doc.pdf (accessed oct. , ). . rdn (resource description network), http://www.rdn .ac.uk/projects/eprints-uk/, (accessed oct. , ); oclc “eprints-uk” ( ), http://www.oclc.org/research/projects/ mswitch/epuk.htm, (accessed oct. , ). . a. macewan, “working with lcsh: the cost of coop- eration and the achievement of access: a perspective from the british library,” presented at the ifla general conference, , http://www.ifla.org/iv/ifla / - e.htm (accessed oct. , ). . ibid.; r. larson, “the decline of subject searching: long- term trends and patterns of index use in an online catalog,” journal of the american society for information science , no. ( ): – . . k. drabenstott et al., “end-user understanding of subject headings in library catalogs,” library resources & technical services , no. (jul. ): – ; bates, “improving user access.” . bates, “improving user access,” (see discussion of entry vocabulary). . ibid. . beat (bibliographic enrichment advisory team, library of congress), “digital tables of contents,” ( ), http://www .loc.gov/catdir/beat/digitoc.html (accessed oct. , ). . d. vizine-goetz, “terminology services, oclc,” ( ), http://www.oclc.org/research/projects/termservices/default .htm (accessed oct. , ). . c. fellbaum, wordnet: an electronic lexical database (cam- bridge, mass.: mit pr., ), http://wordnet.princeton.edu/ (accessed oct. , ); a. csomai, “wordnet bibliography,” ( ). http://lit.csci.unt.edu/~wordnet/ (accessed oct. , ). . bates, “improving user access.” . a. maedche and r. volz, “the ontology extraction and maintenance framework: text-to-onto,” in proceedings of the icdm workshop (san jose, calif.: ieee computer society ( ), http://cui.unige.ch/~hilario/icdm- /dm-km-final/volz .pdf (accessed oct. , ); v. parekh, j. gwo, and t. finin, “mining domain specific texts and glossaries to evaluate and enrich domain ontologies,” in proceedings of the international conference on information and knowledge engineer- ing: ike ‘ (las vegas: csrea press, ), http://ebiquity. umbc.edu/v . /paper/html/id/ / (accessed oct. , ); information technology and libraries | december d. sleeman et al., “enabling services for distributed environ- ments: ontology extraction and knowledge base characteriza- tion,” in proceedings of workshop on knowledge transformation for the semantic web/fifteenth european conference on artificial intelligence (lyon, france: ecai, ), http://www.csd.abdn .ac.uk/~sleeman/published-papers/p -final-ontomine.pdf (accessed oct. , ). ; b. omelayenko, “learning of ontol- ogies for the web: the analysis of existent approaches,” in proceedings of the international workshop on web dynam- ics (london: webdyn, ), http://dcs.bbk.ac.uk/webdyn/ webdynpapers/omelayenko.pdf (accessed oct. , ); r. dhamankar et al., “imap: discovering complex seman- tic matches between database schemas,” in sigmod : proceedings of the acm sigmod international conference on management of data, june – , , paris, france (new york: association for computing machinery, ), http:// www.cs.washington.edu/homes/pedrod/papers/sigmod .pdf (accessed oct. , ); p. cassin et al., “ontology extraction for educational knowledge bases,” lecture notes in computer science, vol. (heidelberg: springer-verlag, ), – ; revised and invited papers from agent-medi- ated knowledge management: international symposium (stanford, calif., mar. – , ), ftp://mas.cs.umass.edu/pub/cassin _ontology-amkm .pdf (accessed oct. , ); t. wang et al., “extracting a domain ontology from linguistic resource based on relatedness measurements,” in the ieee/wic/ acm international conference on web intelligence: proceedings: september – , compiègne university of technology, france (los alamitos, calif.: ieee computer society, ), – , http:// csdl .computer.org/persagen/dlabstoc.jsp?resourcepath=/ dl/proceedings/&toc=comp/proceedings/wi/ / / / toc.xml&doi= . /wi. . (accessed oct. , ). . bates, “improving user access to library catalog and portal information.” . o’neill and chan, “fast: faceted application of subject terminology.” . mason, et al., “infomine: promising directions in vir- tual library development.” salazar information technology and libraries | september author id box for column layout traditional, larger libraries can rely on their physical collection, coffee shops, and study rooms as ways to entice patrons into their library. yet virtual libraries merely have their online presence to attract students to resources. this can only be achieved by providing a fully functional site that is well designed and organized, allowing patrons to navigate and locate information easily. one such technology significantly improving the overall useful- ness of web sites is a content management system (cms). although the cms is not a novel technology per se, it is a tech- nology smaller libraries cannot afford to ignore. in the fall of , the northcentral university electronic learning resources center (elrc), a small, virtual library, moved from a static to a database-driven web site. this article explains the impor- tance of a cms for the virtual or smaller library and describes the methodology used by elrc to complete the project. state of the virtual library the northcentral university elec- tronic learning resource center (elrc), a virtual library, recently moved from a static to a database- driven web site in . before this, the site consisted of static pages and continued to multiply due to the creation and expansion of northcentral university (ncu) pro- grams. to provide the type of service demanded by our internet-savvy patrons, the elrc felt it needed to evolve to the next stage of web man- agement and design. ncu, with a current enrollment of roughly twenty-one hundred full- time students, is one of many for- profit virtual universities (including the university of phoenix, capella, and walden, among others) seeking to carve a niche in the education mar- ket by offering professional degrees entirely online. in the past few years, distance education has experienced exponential growth, causing virtual universities to flourish, but forcing on their libraries the challenge of keeping pace. typically, virtual libraries are manned by a limited staff comprised of one or two librarians who are responsible for all facets of the library, including interlibrary loan, virtual reference, library instruction, and web site management, among other library duties. web site management, as expected, becomes cumbersome when a site exceeds two hundred or more static pages and a clear and structured system is not in place to maintain a proliferating number of web pages. because virtual, for-profit librar- ies do not rely on public funding and taxes, they tend not to be as con- cerned about autonomy as public or state libraries, which must find ways to stay within budget and curtail expenses. on the same note, some academic libraries prefer to maintain a local area network (lan), while other libraries may not have the staff, resources, or need for such a system. thus, for some virtual libraries, such as elrc, the incorporation of tech- nology takes on a more dependent role. that is, where some libraries are encouraged to explore open source applications and create homegrown tools, the virtual, smaller-staffed library finds itself more or less reli- ant on its university’s information technology (it) department. virtual libraries address the needs of distance education students, who demand an equivalent, if not surpass- ing, level of service and instruction as they would expect to find at physical libraries. meeting these needs requires a great deal of creativity, ingenuity, and a strong technical background. recent trends in developing technologies such as mylibrary, learning objects, blogs, virtual chat, and federated searching have broadened the scope of possibilities for the smaller-staffed, virtual library. in particular, a content management system (cms) utilizes a combination of tools that provide numerous advantages, as outlined below: . the creation of templates that maintain a consistent design throughout the site . the convenience of adding, up- dating, and deleting information from a single, online location . the creation and maintenance of interactive pages or learning objects . the implementation of a simple editing interface that eliminates knowledge of extensible hyper- text markup language/hypertext markup language (xhtml/ html) by library staff simply defined, a cms is com- prised of a database, server pages such as active server page (asp), personal home page (php), or coldfusion; a web server—for exam- ple, internet information server (iis), personal web server (pws), or apache; and an editing tool to man- age web content. these resources vary in price, but for a virtual library integrated into a larger university, it is ideal to implement applica- tions and software supported by the university. for the autonomous aca- demic library, this may differ. there are advantages and disadvantages for using proprietary and nonpro- prietary software, and it is left to the library, virtual or physical, to deter- mine the type of resources needed to meet the goals and mission of the university. although the scope of this article focuses on the creation of tools for a homegrown cms, some libraries may wish to explore com- mercial cms packages that include additional services such as technical support. these cms packages will vary in price and services depend- ing on the vendor and the needs of the library. elrc transformed in fall , a group that consisted ed salazar (esalazar@ncu.edu) is reference/web librarian at northcentral university. content management for the virtual library ed salazar article title | author content management for the virtual library | salazar of two librarians, the education chair, and programmer, convened to discuss the redesign of the elrc web site, which had become increas- ingly difficult to manage. specifically, the amount of duplicated content, inconsistent design and layout, and unstructured architecture of the site posed severe navigational and organi- zational problems. the group selected and compared other academic library sites to determine a desired design and theme for the new elrc site. discussions also involved the addi- tion of features such as a site search and breadcrumbs, which the group felt were essential. as a result, the creation of a homegrown cms using proprietary software became the route of choice to meeting the increasing demands of patrons and the need to expand the site. because ncu utilizes microsoft (ms) information system products, it was agreed ms or ms-compatible applications would be used to create the cms, which consisted of sql server, iis, asp, visual basic script (vbscript), jspell iframe, and ms visual interdev. ms visual interdev and jspell iframe supplanted our previous web editor, ms frontpage, which seemed to generate superflu- ous code and thus made it difficult to debug or alter the design and layout of pages. also, using jspell iframe eliminated the need for future ncu librarians to possess an expertise in xhtml/html. with these pieces in place, the arduous task of culling content from static pages and enter- ing it into a database was begun. the database the sql server database helped in organizing and structuring content, and allowed for the creation of tem- plates and administration (admin) pages. in addition, the database played an integral part in creating the search, breadcrumb, and site map features the group so desperately wanted. a significant amount of time was spent weeding the site for infor- mation that had become obsolete or irrelevant to elrc. it should be noted that the group originally attempted to use access for a database but stumbled across several problems, one being the inability to maintain a stable and reliable connection to the database. the templates with the database nearly complete, the programmer began creating asp templates in ms visual interdev. these templates basically serve as the shell of the web page, preserv- ing the design and layout elements of the page while extracting unique content based on a user’s request. in essence, a single template can pro- duce hundreds of pages consistent in design. likewise, a single change to the template can alter the entire design of the site. for the elrc, seven templates were created for more than pages. figure shows the elrc course guides template. figure shows the public view of the elrc course guide template. changes to the templates are done using ms visual interdev, which offers a user-friendly environ- ment for managing web pages. ms visual interdev also includes helpful features, such as highlighting code errors for easy debugging, and the ability to access, create, and maintain stable connections to databases. in addition, the ms visual interdev edi- tor recognizes commonly used asp commands, allowing the user to save time by utilizing keyboard shortcuts when programming. besides creating templates, asp server-include files and cascading style sheets (css) were incorporated, allowing for the easy modification of code on a single file instead of each and every page or template. this, in particular, is time-efficient when having to add or change database connections or design elements. also, the elrc took extra precaution to ensure that style elements met the accessibility requirements and stan- dards set forth by the world wide web consortium (w c), as well as tested the site on other browsers, such as firefox and netscape. as the site continues to grow and expand, so may the need for additional templates. creation or replication of templates is simple, requiring a basic understanding of programming and the re-assigning of new variables in the code to match added or modified tables. there is some speculation in the near future of migrating the site to the asp.net environment for added functionality and security. if and when that time comes, the elrc will be ready. at present, ncu is not considering the use of open source code or applica- tions (the exception being the apache web server); this is primarily due to available technical support, security, and intuitiveness of use associated with commercial software. in addi- tion, the ncu information system was built using commercial software and a complete transition to open source, at the moment, is not possible or desirable. with the templates complete, the elrc began running a prototype of the new site, making it accessible to students and faculty from a link on the old site. a survey was created that allowed users to comment on the new site. one detail of importance to note is that the survey duplicated a prior survey done on the old site in in order to provide the elrc with comparative data. the admin pages the next phase of the project required the creation of admin pages, which would allow content to be quickly added, updated, and deleted on the site. these pages, like the templates, were created in ms visual interdev; display content is housed within the database on the web, thus allowing information technology and libraries | september it to be changed on the fly. figure shows all of the web pages for the elrc within a table. what is particularly convenient about the admin edit pages is the incorporation of the jspell iframe editor, which serves as the front- end editor to the site. the reason for using jspell iframe, as stated earlier, is its ease of use: the simple tool bar provides the basic, essential tools necessary for creating content with- out the daunting number of buttons and menu selections other editors tend to have. also, jspell iframe is reasonably priced and does not entail a complex installation or require any space on local hard drives; instead, the program is maintained on the server. consequentially, all that is required is the insertion of the jspell iframe javascript code into the web pages. in addition to jspell iframe, fields within admin edit pages are or can be pre-populated by content in the data- base. for instance, the title or display order of links can be easily edited or changed. longer text fields comprised of paragraphs are created or modified using jspell iframe. deleting a page is simple, requiring only the click of a delete button on the bottom, right- hand corner. figure shows jspell iframe embedded within an admin edit page. the admin add page is straightfor- ward. information is entered into the fields appearing on a form page, and the proper page type designation is selected from a drop-down menu. yet, more importantly, the admin add and the admin edit pages can filter infor- mation to specific users for security purposes and library needs. figure shows an admin add page. figure shows an admin edit page. the admin pages were designed with flexibility in mind. main col- umn headings may be sorted, as seen in figure , allowing one to locate a particular page. the sorting feature also displays the inner structure of the database that, in turn, identifies parent-child relationships between pages in the elrc, which is useful and necessary when adding pages to the elrc site. due to the careful thought used in creating the admin pages, they have proven to be extremely effective and useful in maintaining a library web site. each and every change to the site can be made on the web, allow- ing content to be edited remotely and eliminating the need for installing and maintaining expensive editing soft- ware on local and remote machines. usability testing with the site completed, the elrc felt it important to perform usability figure . elrc course guide template figure . public view of the elrc course guide template article title | author content management for the virtual library | salazar tests, but how does a virtual library conduct usability testing when all of its students are distance education stu- dents? this is a difficult question that involves some ingenuity to answer. in order to solve this problem, staff members were propositioned (begged) to volunteer for the study. total staff acquired was five. also, a local col- lege class of about ten students was persuaded to participate in the study. granted, the total number of subjects is not representative of the ncu student body; however, substantial changes to the site were made from the data gathered. more usability testing is expected in the immediate future. the findings usability testing complete, the site was launched. during this period, a few minor hang-ups were experienced, including broken links, form page errors, and stray design elements, but these were only minor problems that were quickly fixed. feedback from the elrc survey showed that nearly all of the students and faculty, roughly fifty respondents, approved of the changes by commenting that the site had improved in layout and organi- zation of content as well as naviga- tion. also, responses and comments from usability testing participants were equally positive and encourag- ing. figure shows the new ncu learners elrc home page. although it is difficult to estab- lish a direct connection between the elrc site and usage, recent statistics appear promising. since the inception of the new site in december , the number of visits to the elrc learners home page has jumped percent. this number is expected to rise as ncu continues to grow and students become more acquainted and familiar with the site. the project took nearly six months to complete and required the expertise of a programmer. although program- ming may be outside the requisites of a distance librarian, managing the site is not. a general understanding of control statements and sql is all that is needed. for the distance librarian who spends almost all of his or her time online, these skills can be acquired on the job or by taking introductory programming courses at a local college. in the hope that the site will continue to expand in concert with the growing body of ncu students, recently the elrc added a writing cen- ter and blog. with the entire site now being database driven, adding, updat- ing, deleting content is done effort- lessly. ideally, students and faculty will play a greater role in the development of the elrc site as a result of the changes. involving patrons with the site can play an integral, beneficial role in their academic pursuits. figure . web pages for elrc within a table figure . jspell iframe editor embedded within an admin edit page information technology and libraries | september conclusion the elrc at ncu encourages other virtual or smaller libraries to explore their resources for improving their library web sites, which involves understanding campus resources and personnel. with the ever-burgeoning growth of technological resources, every library—small or large, virtual or physical, public or private—can empower itself to meet the needs of internet-savvy students. it is only a matter of being aware of the resources and putting them to good use. references and notes . the ncu elrc web site is com- prised of three separate sites: the public site www.ncu.edu/elrc (accessed dec. , ), the mentors site http://mentors .ncu.edu/elrc (accessed dec. , ), and the learners site http://learners.ncu .edu/elrc (accessed dec. , ). al- though similar in design, each site is tai- lored to meet the needs of each individual group as well as protect ncu’s resources, services, and information. access to sub- scription resources and personal informa- tion is available upon authentication of the user to the site. . for a detailed overview of virtual libraries, see valerie a. akuna, “virtual universities: the new higher educa- tion paradigm,” estrella mountain col- lege, http://students.estrellamountain .edu/drakuna/virtualuniversities.htm (accessed feb. , ). . u.s. department of education, national center for education statistics, “the condition of education ,” dis- tance education at postsecondary insti- tutions, http://nces.ed.gov/pubsearch/ pubsinfo.asp?pubid= (accessed feb. , ). . for more information on the role of the virtual librarian in a virtual univer- sity, see jan zastrow, “going the distance: academic librarians in the virtual uni- versity,” university of hawaii–kapiolani community college, http://library.kcc .hawaii.edu/~illdoc/de/depaper.htm (accessed jan. , ). . for an overview on developing an open source cms, please see mark dahl, “content management strategy for a college library web site,” information technology and libraries , no. ( ). . for a detailed discussion on dis- tance education and virtual libraries, see smiti gandhi, “academic librarians and distance education: challenges and opportunities,” reference & user services quarterly , no. ( ). . for detailed information on using asp pages for managing databases, see xiaodong li and john paul fullerton, “cre- ate, edit, and manage web database con- tent using active server pages,” library hi tech , no. ( ); see also, bryan h. davidson, “database driven, dynamic content delivery: providing and man- aging access to online resources using microsoft access and active server pages,” oclc systems and services , no. ( ). figure . admin edit page figure . admin add page article title | author content management for the virtual library | salazar . for advantages and disadvantages of open source and proprietary software, see john caroll, “open source versus pro- prietary: both have advantages,” special to cnet asia, http://asia.cnet.com/ builder/program/work/ , , , .htm (accessed feb. , ); see also, stephen shankland, “study: open- source database going mainstream,” cnet, http://ecoustics-cnet.com.com/ study+open-source+databases+going +mainstream/ - _ - .html (accessed feb. , ). . for information on commercial con- tent management vendors and prices, see cms watch, www.cmswatch.com/cms/ vendors (accessed feb. , ). “sql server product overview,” microsoft windows server system, www.microsoft. com/sql/evaluation/overview/default. asp (accessed feb. , ). . for a review on visual interdev, see maggie biggs, “visual studio . demon- strates improved integration,” infoworld , no. ( ), www.infoworld.com/ cgi-bin/displaytc.p ?/reviews/ vstudio .htm (accessed feb. , ). . “checklist of checkpoints for web content accessibility guidelines . ,” w c, www.w .org/tr/wai-webcon tent/full-checklist.html (accessed feb. , ). . jspell iframe , www.jspell .com/iframe-spell-checker.html (accessed dec. , ). figure . elrc learners home page ebsco cover lama cover lita cover index to advertisers cherry information technology and libraries | september article title: subtitle in same font author name and second author author id box for column layout the present study investigated whether there is a cor- relation between user performance and compliance with screen-design guidelines found in the literature. rather than test individual guidelines and their interactions, the authors took a more holistic approach and tested a com- pilation of guidelines. nine bibliographic display formats were scored using a checklist of eighty-six guidelines. twenty-seven participants completed ninety search tasks using the displays in a simulated web environment. none of the correlations indicated that user performance was statistically significantly faster with greater confor- mity to guidelines. in some cases, user performance was actually significantly slower with greater conformity to guidelines. in a supplementary study, a different set of forty-three guidelines and the user performance data from the main study were used. again, none of the correlations indicated that user performance was statistically signifi- cantly faster with greater conformity to guidelines. a ttempts to establish generalizations are ubiquitous in science and in many areas of human endeavor. it is well known that this enterprise can be extremely problematic in both applied and pure science. in the area of human-computer interaction, establishing and evaluating generalizations in the form of interface-design guidelines are pervasive and difficult challenges, particu- larly because of the intractably large number of potential interactions among guidelines. using bibliographic dis- play formats from web catalogs, the present study utilizes global evaluation by correlating user performance in a search task with conformity to a compilation of eighty-six guidelines (divided into four subsets). the literature offers many design guidelines for the user interface, some of which cover all aspects of the user interface, some of which focus on one aspect of the user interface—e.g., screen design. tullis, in chapters in two editions of the handbook of human-computer interaction, reviews the work in this area. the earlier chapter provides a table describing the screen-design guidelines available at that time. he includes, for example, galitz, whom he notes have several hundred guidelines addressing general screen design, and smith and mosier, whom he notes have about three hundred guidelines addressing the display of data. earlier guidelines tended to be generic. more recently, guidelines have been developed for specific applica- tions—e.g., web sites for airline travel agencies, multi- media applications, e-commerce, children, bibliographic displays, and public-information kiosks. although some of the guidelines in the literature are based on empirical evidence, many are based on expert opinion and have not been tested. some of the research- based guidelines have been tested in isolation or in com- bination with only a few other guidelines. the national cancer institute (nci) web site, research-based web design and usability guidelines, rates sixty guidelines on a scale of to based on the strength of the evidence. the more valid the studies that directly support the guideline, the higher the rating. in interpreting the scores, the site advises that scores of , , or suggest that “more evidence is needed to strengthen the designer’s overall confidence in the validity of a guideline.” of the sixty guidelines on the site, forty-six ( . percent) fall into this group. in , the united states department of health and human services web site, research-based web design and usability guidelines, rated guidelines on a different five-point scale. eighty- two guidelines ( . percent) meet the criteria of having strong or medium research support. another forty-eight guidelines ( . percent) are rated as having weak research support. thus, there is some research support for . percent of the guidelines. in addition to the issue of the validity of individual guidelines, there may be interactions among guidelines. an interaction occurs if the effect of a variable depends on the level of another variable—e.g., an interaction occurs if the usefulness of a guideline depends on whether some other guideline is being followed. a more severe problem is the potential for high-order interactions: the nature of a two-way interaction may depend on the level of a third variable, the nature of a three-way interaction may depend on the level of a fourth variable, and so on. because of the combinatorial explosion, if there are more than a few vari- ables the number of possible interactions becomes huge. as cronbach stated: “once we attend to interactions, we enter a hall of mirrors that extends to infinity.” with a large set of guidelines, it is impractical to test all of the guidelines and all of the interactions, including high- order interactions. muter suggested several approaches for handling the problem of intractable high-order interac- tions, including adapting optimizing algorithms such as simplex, seeking “robustness in variation,” re-construing the problem, and pruning the alternative space. the pres- ent study utilizes another approach: global evaluation by joan m. cherry, paul muter, and steve j. szigeti bibliographic displays in web catalogs: does conformity to design guidelines correlate with user performance? joan m. cherry (joan.cherry@utoronto.ca) is a professor in the faculty of information studies; paul muter (muter@psych .utoronto.ca) is an assistant professor in the department of psychology; and steve j. szigeti (szigeti@fis.utoronto.ca) is a doctoral student in the faculty of information studies and the knowledge media design institute, all at the university of toronto, canada. bibliographic displays in web catalogs | cherry, muter, and szigeti correlating user performance with conformity to a set of guidelines. using this method, particular guidelines and interactions are not tested, but the set and subsets are tested globally, and some of the interactions, including high-order interactions, are captured. bibliographic displays were scored using a compilation of guidelines, divided into four subsets, and the performance of users doing a set of search tasks using the displays was measured. an attempt was made to determine whether users find information more quickly on displays that receive high scores on checklists of screen-design guidelines. the authors are aware of only two studies that have investigated conformity with a set of guidelines and user performance, and they both included only ten guide- lines. d’angelo and twining measured the correlation between compliance with a set of ten standards (d’angelo standards) and user comprehension. the d’angelo standards are in the form of principles for web-page design, based on a review of the literature. d’angelo and twining found a small correlation (. ) between number of standards met and user comprehension. they do not report on statistical significance, but from the data provided in the paper it appears that the correlation is not significant. gerhardt-powals compared an interface designed according to ten cognitive engineering principles to two control interfaces and found that the cognitively engineered interface resulted in statistically significantly superior user performance. the guidelines used in the present study were based on a list compiled by chan to evaluate displays of bib- liographic records in online library catalogs. the set of guidelines was broken down into four subsets. participants in this study were given search tasks and clicked on the requested item on a bibliographic display. the main depen- dent variable of interest was response time. ฀ method participants twenty-seven participants were recruited through the university of toronto psychology subject pool. seventeen were female; ten were male. most (twenty) were in the age group to ; three were in the age group to years, and four were in the age group to . one had never used the web; all others reported using the web one or more hours per week. participants received course credit. design to control for the effects of fatigue, practice runs, and the like, the order of trials was determined by two orthogonal x latin squares—one to select a display and one to select a book record. each participant completed five consecutive search tasks—author, title, call number, publisher, and date—in a random order, with each display-book combina- tion. (the order of the five search tasks was randomized each time.) this procedure was repeated, so that in total each participant did ninety tasks ( displays x tasks x repetitions). materials and apparatus the study used nine displays from library catalogs avail- able on the web. they were selected to represent a variety of systems and to illustrate the remarkable diversity in bibliographic displays in web catalogs. the displays dif- fered in the amount of information included, the structure of the display, employment of highlighting techniques, and use of graphical elements. four examples of the nine displays are presented in figures a, b, c, and d. the displays were captured and presented in an interactive environment using active server page (asp) software. the look of the displays was retained, but hypertext links were deactivated. nine different book records were used to provide the content for the displays. items selected were those that would be readily understood by most users—e.g., books by saul bellow, norman mailer, and john updike. the guidelines were based on a list compiled by chan from a review of the literature in human-computer interaction and library science. the list does not include guidelines about the process of design. chan formatted the guidelines as a checklist for bibliographic displays in online catalogs. in work reported in , cherry and cox modified the checklist for use with bibliographic displays in web catalogs. in a paper, cherry reported on evaluations of bibliographic displays in catalogs of aca- demic libraries, based on chan’s data for twelve opacs and data for ten web catalogs evaluated by cherry and cox using a modification of the checklist for web catalogs. the findings showed that, on average, displays in opacs scored percent and displays in web cata- logs scored percent. the checklist of guidelines was modified by herrero-solana and de moya-anegón, who used it to explore the use of multivariate analysis in evaluating twenty-five latin american catalogs. for the present study four questions were removed that were considered less useful from the checklist used in cherry’s analysis. the checklist consisted of four sections or subsets: labels (these identify parts of the bibliographic descrip- tion); text (the display of the bibliographic, holdings/ location, and circulation status information); instructions (includes instructions to users, informational messages, and options available); and layout (includes identifica- tion of the screen, the organization for the bibliographic information technology and libraries | september information, spacing, and consistency of information presentation). items on the checklist were phrased as ques- tions requiring yes/no responses. examples of the items are: labels: “are all fields/variables labeled?” text: “is the text in mixed case (upper and lowercase)?” instructions: “are instructional sentences or phrases simple, concise, clear, and free of typographical errors?” and layout: “is the width of the display no more than forty to sixty characters?” the set used in the present study contained eighty- six guidelines in total, of which forty-eight were generic and could be applied to any application. thirty-eight are specific and apply to bibliographic displays in web catalogs. the experiment was run on a pentium computer with a seventeen-inch sony color monitor with a standard keyboard and mouse. figure a. example of display figure b. example of display figure c. example of display figure d. example of display bibliographic displays in web catalogs | cherry, muter, and szigeti procedure participants were tested individually. five practice trials with a display and book record not used in the experi- ment familiarized the participant with the tasks and software. at the beginning of a trial, the message “when ready, click” appeared on the screen. when the participant clicked on the mouse, a bibliographic display appeared along with a message at the top of the screen indicating whether the participant should click on the author, title, call number, publisher, or date of publication—e.g., “current task: author.” participants clicked on what they thought was the correct answer. if they clicked on any other area, the display was shown again. an incorrect click was not defined as an error—in effect, percent correct was always —but an incorrect click would of course add to the response time. the software recorded the time to suc- cessfully complete each search, the identification for the display and the book record, and the search-task type. when a participant completed the five search tasks for a display, a message was shown indicating the average response time on that set of tasks. when participants completed the ninety search tasks, they were asked to rank the nine displays according to their preference. for this task, a set of laminated color printouts of the displays was provided. participants ranked the displays, assigning a rank of to the display that they preferred most, and to the one they preferred least. they were also asked to complete a short background questionnaire. the entire session took less than forty-five minutes. scoring the displays on screen design guidelines the authors’ experience has indicated that judging whether a guideline is met can be problematic: evalua- tors sometimes differ in their judgments. in this study, three evaluators assessed each of the nine displays inde- pendently. if there was any disagreement amongst the evaluators’ responses for a given question for a given display, that question was not used in the computation of the percentage score for that display. (a guideline regard- ing screen density was evaluated by only one evaluator because it was very time-consuming.) the total number of questions used to assess each display was eighty-six. the number of questions on which the evaluators disagreed ranged from twelve to thirty across the nine displays. all questions on which the three evaluators agreed for a given display were used in the calculation of the percent- age score for that display. hence the percentage scores for the displays are based on a variable set and number of questions—from fifty-six to seventy-four. the subset of questions on which the three evaluators agreed for all nine displays was small—twenty-two questions. ฀ results with regard to conformity to the guidelines, in addition to the overall scores for each display, which ranged from percent to percent, the percentage score was calculated for each subset of the checklist (labels, text, instructions, and layout). the time to successfully complete each search task was recorded to the nearest millisecond. (for some unknown reason, six of the , response times recorded [ x ] were milliseconds. the program was written in such a way that the response-time buffer was cleared at the time of stimulus presentation, in case the participant clicked just before this time. these trials were treated as missing values in the calculation of the means.) six mean response times were calculated: author, title, call number, publisher, date, and the sum of the five response times, called all tasks. the mean of all tasks response times ranged from , milliseconds to , milliseconds for the nine formats. the nine display formats differed significantly on this variable according to an analysis of variance, f( , ) = . , p < . . the correlations between response times and guide- lines-conformance scores are presented in table . it is important to note that a high correlation between response time and conformity to guidelines indicates a low correla- tion between user performance (speed) and conformity to guidelines. row of table contains correlations between the total guidelines score and response times; column contains correlations between all tasks (the sum of the five response times) and guidelines scores. of course, the cor- relations in table are not all independent of each other. only five of the thirty correlations in table are signifi- cant at the . level, and they all indicate slower response times with higher conformity to guidelines. of the six correlations in table indicating faster response times with higher conformity to guidelines, none approaches statistical significance. the upper left-hand cell of table indicates that the overall correlation between total scores on the guidelines and the mean response time across all search tasks (all tasks) was . (df = , p = . )—i.e., conformity to the overall checklist was correlated with slower overall response times, though this correlation did not approach statistical significance. figure shows a scatter plot of the main independent variable, overall score on the checklist of guidelines, and the main dependent variable, the sum of the response times for the five tasks (all tasks). figure shows a scatter plot for the highest obtained correlation: between score on the overall checklist of guidelines and the time to complete the title search task. visual inspection suggests patterns con- sistent with table : no correlation in figure , and slower search times with higher guidelines scores in figure . finally, correlations were computed between prefer- ence and response times (all tasks response times and five information technology and libraries | september specific-task response times) and between preference and conformity to guidelines (over- all guidelines four subsets of guidelines). none of the eleven correlations approached statisti- cal significance. ฀ supplementary study to further validate the results of the main study, it was decided to score the interfaces against a different set of guidelines based on the u.s. department of health and human services research-based web design and usability guidelines. this set consists of guidelines and includes a rating for each guide- line based on strength of research evidence for that guide- line. the present study started with eighty-two guidelines that were rated as having either moderate or strong research support, as the definitions of both of these include “cumulative research-based evidence.” compliance with guidelines that address the process of design can only be judged during the design process, or via access to the interface designers. since this review process did not allow for that, a total of nine process-focused guidelines were dis- carded. this set of seventy-three guidelines was then com- pared with the sixty-guideline nci set, research-based web design and usability guidelines, intending to add any outstanding nci guidelines supported by strong research evidence to the existing list of seventy-three. however, all of the strongly supported nci guidelines were already represented in the original seventy-three. finally, the guidelines in the iso , ergonomic requirements for office work with visual display terminals (vdts), part (guidance on usability), part (presentation of information ), and part (menu dialogues ) were compared to the existing set of seventy-three, with the intention that any prescriptive guideline in the iso set that was not already included in the original seventy-three would be added. again, there were none. the seventy-three guidelines were organized into three thematic groups: ( ) layout (the organization of textual and graphic material on the screen), ( ) interaction (which included navigation or any element with which the user would interact), and ( ) text and readability. all of the guidelines used were written in a manner allowing readers room for interpretation. the authors explicitly stated that they were not writing rules, but rather, guidelines, and recognized that their application must allow for a level of flexibility. this ambiguity creates problems in terms of assessing displays. in this study, two evaluators independently assessed the nine displays. the first evaluator applied all seventy-three guidelines and found thirty to be nonapplicable to the specific types of interfaces considered. the second evaluator applied the shortened list of forty-three guidelines. following the independent evaluations, the two evaluators compared assessments. the initial rate of agreement between the two assessments ranged from percent to percent across the nine displays. in cases where there was disagreement, the evaluators discussed their rationale for the assessment in order to achieve consensus. ฀ results of supplementary study as with the initial study, in addition to the overall scores for each display, the percentage score was calculated for each subset of the checklist (labels, interaction, and text and readability). it is worth noting that the overall scores witnessed higher compliance to this second set of guide- lines, ranging from percent to percent. the correla- tions between response times and guidelines-conformance scores are presented in table . again, it is important to note that a high correlation between response time and confor- mity to guidelines indicates a low correlation between user performance (speed) and conformity to guidelines. row of table contains correlations between the total guidelines score and response times; column contains correlations between all tasks (the sum of the five response times) and guidelines scores. of course, the correlations in table are not all independent of each other. only one of the twenty-four correlations in table table . correlations between scores on the checklist of screen design guidelines and time to complete search tasks: pearson correlation (sig. - -tailed); n= all cells all tasks author title call # publisher year total score: . (. ) . (. ) . (. ) . (. ) . (. ) . (. ) labels: . (. ) . (. ) . (. ) . (. ) . (. ) . (. ) text: -. (. ) -. (. ) . (. ) -. (. ) -. (. ) -. (. ) instructions: . (. ) . (. ) . (. ) . (. ) . (. ) . (. ) layout: . (. -. (. ) . (. ) . (. ) . (. ) . (. ) bibliographic displays in web catalogs | cherry, muter, and szigeti is significant at the . level, and it indicates a slower response time with higher conformity to guidelines. of the ten correlations in table indicating faster response times with higher conformity to guidelines, none approaches statistical significance. the upper left-hand cell of table indicates that the overall correlation between total scores on the guidelines and the mean response time across all search tasks (all tasks) was . (p = . )—i.e., conformity to the overall checklist was correlated with slower overall response times, though this correlation did not approach statistical significance. figure shows a scatter plot of the main independent variable, overall score on the checklist of guidelines, and the main dependent variable, the sum of the response times for the five tasks (all tasks). figure shows a scatter plot for the highest-obtained correlation: between score on the text and readability category of guidelines and the time to complete the title search task. visual inspection suggests patterns consistent with table : no correlation in figure , and slower search times with higher guidelines scores in figure . ฀ discussion in the present experiment and the supplementary study, none of the correlations indicating faster user perfor- mance with greater conformity to guidelines approached statistical significance. in some cases, user performance was actually significantly slower with greater conformity to guidelines—i.e., in some cases, there was a negative correlation between user performance and conformity to guidelines. the authors are aware of no other study indicating a negative correlation between user performance and con- formity to interface design guidelines. some researchers would not be surprised at a finding of zero correlation between user performance and conformity to guide- lines, but a negative correlation is somewhat puzzling. a negative correlation implies that there is something wrong somewhere—perhaps incorrect underlying theories or an incorrect body of assumptions. such a negative correla- tion is not without precedent in applied science. in the field of medicine, before the turn of the twentieth century, seeing a doctor actually decreased the chances of improv- ing health. presumably, medical guidelines of the time were negatively correlated with successful practice, and the negative correlation implies not just worthlessness, but medical theories or beliefs that were actually incorrect and harmful. the boundary conditions of the present findings are unknown. the present findings may be specific to the tasks employed—fairly simple search tasks. the findings may apply only to situations in which the user is switch- ing formats frequently, as opposed to situations in which each user is using only one format. (a between-subjects design would test this possibility.) the findings may be specific to the two sets of guidelines used. with sets of ten guidelines, d’angelo and twining and gerhardt-powals found positive correlations between user performance and conformity to guidelines (though apparently not statisti- cally significantly in the former study). the guidelines used in the authors’ main study and supplementary study tended to be more detailed than in the other two studies. detailed guidelines are sometimes seen as advantageous, since developers who use guidelines need to be able to interpret the guidelines in order to implement them. however, perhaps following a large number of detailed figure . scatter plot for overall score on checklist of screen design guidelines and time to complete set of five search tasks figure . scatter plot for overall score on checklist of screen design guidelines and time to complete “title” search tasks information technology and libraries | september guidelines reduces the amount of personal judgment used and results in less effective designs. (designers of the nine displays used in the present study would not have been using either of the sets of guidelines used in our studies but may have been using some of the sources from which our guidelines were extracted.) as noted by cheepen in discussing guidelines for voice dialogues, sometimes a designer’s experience may be more valuable than a par- ticular guideline. the lack of agreement in interpreting the guide- lines was an unexpected but interesting factor re- vealed during the collec- tion of data in both the main study and the sup- plementary study. while a higher rate of agree- ment had been expected, the differences raised an important point in the use of guidelines. if guidelines intentionally leave room for interpretation, what factor does expert opinion and experience play in design? in the main study, the number of guidelines on which the evaluators disagreed ranged from percent to percent across the nine displays. in the supplemen- tary study, both evaluators had experience in interface design through a number of different roles in the design process (both academic and professional). this meant the evaluators’ interpretations of the guidelines were informed by previous experience. the initial level of disagreement ranged from percent to percent across the nine dis- plays. while it was possible to quickly reach consensus table . correlations between scores on subset of the u.s. dept. of health and human services ( ) research–based web design and usability guidelines and time to complete search tasks: pearson correlation (sig. - -tailed); n= all cells all tasks author title call # publisher year total score: . (. ) . (. ) . (. ) -. (. ) . (. ) . (. ) layout: -. (. ) -. (. ) -. (. ) -. (. ) . (. ) -. (. ) text: . (. ) -. (. ) . (. ) -. (. ) -. (. ) -. (. ) interaction: . (. ) . (. ) . (. ) . (. ) . (. ) . (. ) figure . scatter plot for subset of u.s. department of health and human services ( ) research–based web design and usability guidelines conformance score and total time to com- plete five search tasks figure . scatter plot for text and readability category of u.s. department of health and human services ( ) research–based web design and usability guidelines and time to complete “title” search tasks bibliographic displays in web catalogs | cherry, muter, and szigeti on a number of assessments (because both evaluators recognized the high degree of subjectivity that is involved in design), it also led to longer discussions regarding the intentions of the guideline authors. a majority of the differ- ences involved lack of guideline clarity (where one evalu- ator had indicated a meet-or-fail score, while another felt the guideline was either unclear or not applicable). does this imply that guidelines can best be applied by commit- tees or groups of designers? the dynamic of such groups would add another complex variable to understanding the relationship between guideline conformity and user performance. future research should test other tasks and other sets of guidelines to confirm or refute the findings of the present study. there should also be investigation of other potential predictors of display effectiveness. for example, would the ratings of usability experts or graphic designers for a set of bibliographic displays be positively correlated with user performance? crawford, in response to a paper presenting findings from an evaluation of bibliographic displays using a previous version of the checklist of guidelines used in the main study, commented that the design of bibliographic displays still reflects art, not science. several researchers have discussed aesthetics and user interface design. reed et al. noted the need to extend our understanding of the role of aesthetic elements in the context of user-interface guidelines and standards. ngo, teo, and byrne dis- cussed fourteen aesthetic measures for graphic displays. norman discussed these ideas in “emotions and design: attractive things work better.” tractinsky, katz, and ikar found strong correlations between perceived aesthetic appeal and perceived usability. most empirical studies of guidelines have looked at one variable only or, at the most, a small number of variables. the opposite extreme would be to do a study that exam- ines a large number of variables factorially. for example, assuming eighty-six yes/no guidelines for bibliographic displays, it would be theoretically possible to do a factorial experiment testing all possible combinations of yes/no— to the th power. in such an experiment, all two-way interactions and higher interactions could be assessed, but such an experiment is not feasible. what the authors have done is somewhere between these two extremes. this study has the disadvantage that we cannot say anything about any individual guideline, but it has the advantage that it captures some of the interactions, including high- order interactions. despite the present results, the authors are not recom- mending abandoning the search for guidelines in interface design. at a minimum, the use of guidelines may increase consistency across interfaces, which may be helpful. however, in some research domains, particularly when huge numbers of potential interactions result in extreme complexity, it may be advisable to allocate resources to means other than attempting to establish guidelines, such as expert review, relying on tradition, letting natural selection take its course, utilizing the intuitions of design- ers, and observing user-interaction. indeed, in pure and applied research in general, perhaps more resources should be allocated to means other than searching for explicit generalizations. future research may better indi- cate when to attempt to establish generalizations and when to use other methods. ฀ acknowledgements this work was supported by a social sciences and humanities research council general research grant awarded by the faculty of information studies, university of toronto, and by the natural sciences and engineering research council of canada. the authors wish to thank mark dykeman and gerry oxford who developed the software for the experiment; donna chan, joan bartlett, and margaret english, who scored the displays with the first set of guidelines; everton lewis, who conducted the experimental sessions; m. max evans, who helped score the displays with the supplementary set of guidelines; and robert l. duchnicky, jonathan l. freedman, bruce oddson, tarjin rahman, and paul w. smith for helpful comments. references and notes . see, for example, a. chapanis, “some generalizations about generalization,” human factors , no. ( ): – . . t. s. tullis, “screen design,” in handbook of human-com- puter interaction, ed. m. helander (amsterdam: elsevier, ), – ; t. s. tullis, “screen design,” in handbook of human- computer interaction, d ed., eds. m. helander, t. k. landauer, and p. prabhu (amsterdam: elsevier, ), – . . w. o. galitz, handbook of screen format design, d ed. (wellesley hills, mass.: qed information sciences, ); s. l. smith and j. n. mosier, guidelines for designing user interface software, technical report esd-tr- - (hanscom air force base, mass.: usaf electronic systems division, ). . c. chariton and m. choi, “user interface guidelines for enhancing the usability of airline travel agency e-com- merce web sites,” chi ‘ extended abstracts on human fac- tors in computing systems, apr. – , (minneapolis, minn.: acm press), – , http://portal.acm.org/citation .cfm?doid= . (accessed dec. , ); m. g. wad- low, “the andrew system; the role of human interface guidelines in the design of multimedia applications,” current psychology: research and reviews (summer ): – ; j. kim and j. lee, “critical design factors for successful e-commerce systems,” behaviour and information technology , no. ( ): – ; s. giltuz and j. nielsen, usability of web sites for children: information technology and libraries | september design guidelines (fremont, calif.: nielsen norman group, ); juliana chan, “evaluation of formats used to display bibliographic records in opacs in canadian academic and public libraries,” master of information science research proj- ect report (university of toronto: faculty of information stud- ies, ); m. c. maquire, “a review of user-interface design guidelines for public information kiosk systems,” international journal of human-computer studies , no. ( ): – . . national cancer institute, research-based web design and usability guidelines ( ), www.usability.gov/guidelines/index .html (accessed dec. , ). . u.s. department of health and human services, research- based web design and usability guidelines ( ), http://usability .gov/pdfs/guidelines.html (accessed dec. , ). . l. j. cronbach, “beyond the two disciplines of scientific psychology,” american psychologist , no. ( ): – . . p. muter, “interface design and optimization of read- ing of continuous text,” in cognitive aspects of electronic text processing, eds. h. van oostendorp and s. de mul (norwood, n.j.: ablex, ), – ; j. a. nelder and r. mead, “a sim- plex method for function minimization,” computer journal , no. ( ): – ; t. k. landauer, “research methods in human-computer interaction,” in handbook of human-computer interaction, ed. m. helander (amsterdam: elsevier, ), – ; r. n. shepard, “toward a universal law of generalization for psychological science,” science (sept. , ): – . . j. d. d’angelo and j. twining, “comprehension by clicks: d’angelo standards for web page design, and time, compre- hension, and preference,” information technology and libraries , no. ( ): – . . j. d. d’angelo and s. k. little, “successful web pages: what are they and do they exist?” information technology and libraries , no. ( ): – . . d’angelo and twining, “comprehension by clicks.” . j. gerhardt-powals, “cognitive engineering principles for enhancing human-computer performance,” international journal of human-computer interaction , no. ( ): – . . chan, “evaluation of formats.” . ibid. . joan m. cherry and joseph p. cox, “world wide web displays of bibliographic records: an evaluation,” proceed- ings of the th annual conference of the canadian association for information science (toronto, ontario: canadian association for information science, ), – . . joan m. cherry, “bibliographic displays in opacs and web catalogs: how well do they comply with display guide- lines?” information technology and libraries , no. ( ): – ; cherry and cox, “world wide web displays of bibliographic records.” . v. herrero-solana and f. de moya-anegón, “bibliograph- ic displays of web-based opacs: multivariate analysis applied to latin-american catalogs,” libri (june ): – . . u.s. department of health and human services, research- based web design and usability guidelines, xxi. . international organization for standardization, iso - : ergonomic requirements for office work with visual display terminals (vdts)—part : guidance on usability (geneva, swit- zerland: international organization for standardization, ); international organization for standardization, iso - : ergonomic requirements for office work with visual display termi- nals (vdts)—part : presentation of information (geneva, swit- zerland: international organization for standardization, ); international organization for standardization, iso - : ergonomic requirements for office work with visual display ter- minals (vdts)—part : menu dialogues (geneva, switzerland: international organization for standardization, ). . u.s. department of health and human services, research- based web design and usability guidelines. . ivan illich, limits to medicine: medical nemesis: the expro- priation of health (harmondsworth, n.y.: penguin, ). . d’angelo and twining, “comprehension by clicks”; ger- hardt-powals, “cognitive engineering principles.” . c. cheepen, “guidelines for dialogue design—what is our approach? working design guidelines for advanced voice dia- logues project. paper ,” ( ), www.soc.surrey.ac.uk/research/ reports/voice-dialogues/wp .html (accessed dec. , ). . w. crawford, “webcats and checklists: some caution- ary notes,” information technology and libraries , no. , ( ): – ; cherry, “bibliographic displays in opacs and web catalogs.” . p. reed et al., “user interface guidelines and standards: progress, issues, and prospects,” interacting with computers , no. ( ): – . . d. c. l. ngo, l. s. teo, and j. g. byrne, “formalizing guidelines for the design of screen layouts,” displays , no. ( ): – . . d. a. norman, “emotion and design: attractive things work better,” interactions , no. ( ): – . . n. tractinsky, a. s. katz, d. ikar, “what is beautiful is usable,” interacting with computers , no. ( ): – . guest editorial | hirst o rganization structure and reorganization are never exciting topics. the world rarely pauses to take a deep breath or offer a round of applause when an organization adds a new committee or decides to split into subgroups. however, organizations frequently inform the patterns and processes of change—as well as no change. recently, the ex libris users of north america (eluna) group reorganized. processes and outcomes were similar to those i observed many years before when the library information and technology association (lita) restructured, and i labeled the process litaish. john webb subsequently asked me to elaborate through an information technologies and libraries (ital) editorial. ■ lita—an organizational recap in , lita launched a bold reorganization. sections and committees were abolished and a new structure, the inter- est group, was created with the hope of significant benefits to the organization. the final report of the long-range plan implementation committee of may , , stated: the main thrust of the reorganization . . . was the estab- lishment and encouragement of interest groups, which were intended to reflect topics of current interest to members and to have a structure which allows for easy creation and easy elimination as interests and technology change. interest groups could be formed . . . from as few as ten lita members and were empowered to plan and present programs, institutes, and preconferences . . . linda knutson, who became executive director of lita in february , “has . . . been impressed by the increase in the level of participation and by the tremendous energy that the players have; they want to contribute, and they plunge in with both feet.” these comments are from conversations with linda knutson quoted in “lita’s first twenty-five years: a brief history,” by stephen r. salmon in the march silver anniversary issue of ital. twenty years later, the lita organization and, specifi- cally, the lita interest groups (igs) continue to provide forums for discussion, create conference programs, insti- tutes, and preconferences. the igs hold the content of the organization with minimal administrative overhead, irregular leadership, and virtually no bylaws. ■ naaug—the deconstruction of a classic model aleph, the ex libris integrated library solution (ils) software, is used in numerous countries. the north american aleph user’s group (naaug) existed from to . the organization had a reasonably classic structure with a steering committee and ad hoc groups to work on annual software enhancements, focus groups, and conference planning. the organization was very cen- tralized with all appointments to subgroups made by the steering committee. developments outside the ils put pressure on naaug to reorganize. ex libris was offering numerous new products, some of which complemented, some of which were independent of the ils. as with any organization, there was some pressure to retain all or part of the status quo from those who were hesitant to change or change radically. leaders, including myself, were cautious, always questioning whether new developments would work and be effective. ■ eluna emerges the new ex libris users’ organization, eluna, is com- posed of the steering committee, product groups (pgs), and interest groups (igs). i was intrigued with the formation of eluna igs and believe that this structure was an offspring of the lita igs. the eluna igs have very little bureaucracy to hinder the creativity and energy that lita wanted to capture. there is no minimum number of participants in an eluna ig, the creation of which can be proposed by any single individual. each group must write a brief annual report, have a contact person whose name and e-mail is posted on the web site, and may have an optional electronic discussion list. the groups can meet at the annual conference or anywhere they choose and a virtual ig is not discouraged. the igs may get involved in product enhancements, but it is fine to leave this work to the pgs. currently, igs are organized around such areas as function, type of library, and particular software. some examples: ■ data representation (special scripts) ■ law ■ edi ■ music ■ government publications ■ shared systems (consortia) ■ ill ■ sql guest editorial: organizational structure— yesterday informs the present donna hirst donna l. hirst (donna-hirst@uiowa.edu) is project coordinator, library information technology, university of iowa libraries, and a member of the ital editorial board. information technology and libraries | december ■ large research libraries ■ z . ■ what happens next? the eluna structures of steering committee, pgs, and igs are off to a good start. because each of these is empowered to work independently, a communication matrix needs to be put into place so that all interested or affected parties are adequately informed. in the future, a process will need to be created to iden- tify groups that need to be disbanded. lita solved this problem with the periodic renewal process. in eluna, the contact person may be able to assume this responsibility. we live in an age where “opening” offers a context for change. opening implies new possibilities and few restric- tions. open systems . . . open access . . . open source. it appears to me that eluna is continuing a tradition that lita began twenty-five years ago with an open organiza- tion. put people into a group, stir lightly, and watch what comes out of the pot. ■ academic web site design and academic templates | peterson academic web site design continues to evolve as colleges and universities are under increasing pressure to create a web site that is both hip and professional looking. many colleges and universities are using templates to unify the look and feel of their web sites. where does the library web site fit into a comprehensive campus design scheme? the library web site is unique due to the wide range of services and content available. based on a poster session presented at the twelfth annual association of college and research libraries conference in minneapolis, minnesota, april , this paper explores the preva- lence of university-wide academic templates on library web sites and discusses factors libraries should consider in the future. c ollege and universities have a long history with the web. in the early s, university web sites began as piecemeal projects with varying degrees of complexity—many started as informational sites for various technologically advanced departments on campus. over the last decade, these web sites have become a vital part of postsecondary institutions and one of their most visible faces. academic web sites communicate the brand and mission of an institution. they are used by prospective students to learn about an institution and then used later to apply. current students use them to pay tuition bills, register for classes, access course materials, participate in class discussions, take tests, get grades, and more. online learning and course-management software programs, such as blackboard, continue to increase the use of web sites. they are now an important learning tool for the entire campus community and the primary communication tool for current students, parents, alumni, the community, donors, and funding organizations. web site standards have developed since the s. usability and accessibility are now important tenets for web site designers, especially for educational institutions. as a result, campus web designers or outside consultants are often responsible for designing large parts of the academic web site. as web sites have grown, ongoing maintenance is an important workload issue. databases and other technologies are used to simplify daily updates and changes to web sites. this is where the academic template fits in. an academic template can be defined as a common or shared template used to control the formatting of web pages in different departments on a campus. generally, administrators will mandate the use of a specific template or group of templates. this mandate includes guidelines for such things as layout, design, color, font, graphics, and navigation links to be used on all web pages. often, the templates are administered using content management systems (cmss) or web development software such as macromedia’s contribute. these programs give different levels of editing rights to individuals, thus keeping tight control over particular web pages or even parts of web pages. academic templates give the web site administra- tor the ability to change the template and update all pages with a single keystroke. for example, the web site administrator may give edit- ing rights to content editors, such as librarians, to edit only the center section of the web page. the remaining parts of the page such as the top, sides, and bottom are locked and cannot be edited. the result of using templates is that the university web site is very unified and consistent. this is particularly important in creating a brand for the university. well-branded institutions have the opportunity to increase revenue, improve administration and faculty staffing, improve retention, and increase alumni relation- ships. but what about the library? libraries are one of the most visited web pages on a university’s web site. thus, the design of the library page can be crucial to a well-designed academic web site. the library web site can set a tone for an institution and help prospective students get a feel for the campus. belanger, mount, and wilson contend it is important for the image of an institution to match the reality. if there is discord between the two, students may choose an inappropriate college and quickly drop out, lowering a campus’s reten- tion data. the library web site can also be important in the recruitment of new faculty members. in addition, libraries use their web sites for marketing, public relations, and fund-raising for the library. library web sites are crucial to delivering data, research tools, and instruction to students, faculty, staff, and com- munity patrons. more than percent of students access the library from their home computers, and percent prefer this form of access. today, the web site connects users with article citations and databases, library catalogs, full-text journals, magazines, newspapers, books, videos, dvds, e-books, encyclopedias, streaming music and video, and more. users access subject-specific research guides, library tutorials, information-literacy instruction, and critical evaluation tools. services such as interlibrary loan (ill), reference management programs such as endnote or refworks, and print and electronic reserves are also used via the web. users get help with doing research by e-mail and virtual chat. in addition, libraries are digital repositories for a growing number of digital historic docu- ments and archives. academic web site design and academic templates: where does the library fit in? kate peterson kate peterson (katepeterson@gmail.com) is an information literacy librarian at capella university, minneapolis, minnesota. information technology and libraries | december how common are academic templates in library web sites? what effect do they have on the content and services provided by libraries? ■ methods for the purposes of this study, a list of doctoral, master’s, and bachelor of arts (ba) institutions (private and public) based on the carnegie classification of institutions of higher education was created and a random number table was used to select a sample of web pages (n= ). home pages, admissions pages, departmental pages, and library web pages were analyzed. a similarly sized sample of each type was selected to give a broad overview of trends— percent of doctoral institutions (n= ), percent of mas- ter’s institutions (n= ), and percent of ba institutions (n= ). the following questions were asked: ■ does the college or university web site use an aca- demic template? ■ if yes, is the library using the template, and for how much of the library web site? ■ to what extent is the template being used? primarily, a web site was determined to be using an aca- demic template based on the look of the site. for example, if the majority of the web elements (top banner, navigation) all matched, then the web site was counted as using some sort of template. use and nonuse of content management system (cms) software behind the web site was not con- sidered in this study—only the look of the web site. ■ results a majority of college and university web sites ( percent) use an academic template. fifty percent of the librar- ies surveyed use the academic template for at least the library’s home page. of that number, about percent of libraries use the template on a majority of the library pages. roughly percent of the total libraries surveyed did not use the academic template, and approximately percent of academic web sites do not use any sort of uni- fied academic template. smaller ba institutions are more likely to use the academic template on multiple library pages than doctoral institutions, which tend to have their own library design or template (see table ). for those libraries that did not use the academic tem- plate on every library page, the most commonly used elements template were the top header (which often has the university seal or an image of the university), the top navigation bar (with university-wide links), and the bot- tom footer, which often contains the university address, privacy statement, or legal disclaimers. less frequently used elements were the bottom navigation bar, and the left or right navigation bar with university-wide links (see tables – ). ■ discussion while many colleges and universities use academic templates, only about half of their libraries follow suit. libraries using the template often use selected parts of the template, or only use the template on their home page. though not considered in this study, there may be a correlation between institution size and template use, as larger institutions are more likely to have library web designers and thus use the academic template only on the library’s home page. while academic templates can cause libraries many problems, there are also many benefits to be considered. ■ problems with academictemplates on library web sites the primary concern with any template is how much space is available for content. for example, there may be a very small box for the page content while images, banner bars, and large navigation links may take up most of the real estate on the page. this problem can be exacerbated for libraries because there are so many different types of content such as the library catalog, databases, tutorials, forms, ill, and other library services delivered via the web. libraries can be caught between the design imposed by the academic template and the rigid size requirements from outside vendors such as database companies, ill or reserve modules, federated search products, or others. academic templates are usually mandated by admin- istrators without a full understanding of the specific content and uses of the library web site. many problems can occur when trying to fit an existing library web site into a poorly designed academic template. it can be very difficult to modify the template effectively for the library’s purposes. an example of one specific problem is confus- ing links on the template, where a link on every page to the “university catalog” links to the course catalog and not the library catalog, which is very confusing for users. another example is a search box as part of the academic template—what are users searching? the university web site? the library web site? the library catalog? the world wide web? another drawback to using academic templates for library web sites can be the time involved in training librar- ians, staff, and library web site administrators. the existing academic web site design and academic templates | peterson content must be fit into the new template—a huge project, given that many library web sites contain one thousand pages or more. generally, a decision to use a template is accompanied by a decision to use a cms or new web-page editor. this takes yet more time to train individuals on the new software in addition to the new template. ■ benefits of using academic templates one of the benefits for libraries using an academic tem- plate is the ability to exploit the expertise of the web site designers who created the template. the academic template often incorporates images, logos, and brand- ing that the library may not be able to design otherwise. many libraries do not have professional web designers on staff; even if they do, there often is no one person who designs and maintains the entire library web site. instead, different parts of a library web site are designed and maintained by different individuals with varying degrees of web site ability. as a result, many library web sites are a mix of styles, which can be disorienting for students who are familiar with the university’s “look.” web site uniformity has a positive effect on usability since familiar- ity with one part of the web site helps students, faculty, table . percentages of occurrences of academic templates no academic template (%) library not using template (%) library using template—transition or top page (%) library using template—majority of pages (%) bachelor of arts master’s doctoral table . occurrence of templates in academic and library web sites no academic template library not using template library using template— transition or top page library using template— majority of pages total sites analyzed bachelor of arts master’s doctoral total table . percentages of occurrence for institutions using the academic-wide template for first page of library web site or libraries using modified academic template ba (%) master’s (%) doctoral (%) all colleges and universities (%) top header (no navigation) top navigation bottom header (no navigation) bottom navigation left navigation right navigation information technology and libraries | december and staff navigate other parts of the web site. even web site basics such as knowing the color and style of the links and how to navigate to different pages can be helpful. another benefit is academic templates are generally ada compliant as required under section of the rehabilitation act of . as usability and usability test- ing become more prevalent, academic template designers may also test the template and navigation for usability. such testing will improve the template and thus the library web site as well. ■ trends in academic and library web sites colleges and universities are responding to a new genera- tion of students, the majority of whom have grown up with computers. in trying to meet their needs and desires, many academic web sites have high-quality photographs, quotes, and testimonials from the universities’ students on their home pages. more and more materials are being placed online to allow both prospective and current stu- dents to do what they need to do twenty-four hours a day, from registering for classes to handing in research papers. many web sites have interactive elements such as instant polls or quizzlets or use instant messaging to connect with tech-savvy students. for example, prospective students can chat with admissions staff members or current students about what it is like to attend a particular university. a large number of sites also highlight weblogs written by current students or those studying abroad. these features allow students to use the technology they are comfortable with to maximize their academic experience. numerous library web sites are changing as well, fea- turing a library catalog, article database, or federated search box on the home page to allow users to search instantly. additionally, library sites are beginning to include images of students using the library, external or internal shots of the building, flash graphics, icons, and sound. many incorporate screen captures to help users navigate specific databases or forms. in addition, an increasing number of libraries use weblogs to give more of a dynamic quality with daily library news and announcements. ■ strategies for using academic templates based on comments received in april during the poster session, and in recent electronic discussion list postings, many academic libraries are dealing with these issues. libraries should work on creating a mission state- ment and objectives for their web sites that expand upon the library’s mission, the institutional web site’s mission, and the institution’s overall mission and brand. librarians must be knowledgeable about web site usability and trends in web site design in order to communicate effectively to designers and administrators. librarians should also become members of campus web committees and be a voice for library users during the design process. teaching administrators and campus web designers about the library and the library web site’s prominence are important tools to successfully deal with any proposed university-wide academic templates. for example, a librar- ian could mock-up a few pages, conduct informal usability testing, and invite administrators to learn firsthand about potential problems library users could experience with a template. librarians could also propose a modified template that uses a few key elements from the academic template. this would maintain the brand but retain enough space for important library content. connecting with other librarians and learning from each other’s successes and failures will also help bring insight into this academic template issue. ■ conclusion the use of academic templates is only going to increase as institutional web sites grow in complexity and importance. libraries are an important part of institutions both physi- cally—on campus—and virtually—as part of the campus web site. academic templates are part of a unified design scheme for colleges and universities. librarians must work with both library and university administrators to create a well-designed but usable library web site. they must advocate for library users and continue to help students and faculty access the rich resources and services available from the library. library administrators need to allocate resources and staff time to improve their web sites and to work in concert with academic web site designers to merge the best of the academic template to the best of the library site while not sacrificing users’ needs. the result will be highly used, highly usable library web sites that attract students and keep them coming back to access the fantastic world of information available in today’s academic libraries. ■ references . robert sevier, “university branding: keys to success,” university business , no. ( ): – . . mignon adams and richard m. dougherty, “how useful is your homepage? a quick and practical approach to evaluat- ing a library’s web site,” college & research libraries news , no. ( ): – . academic web site design and academic templates | peterson . charles belanger, joan mount, and mathew wilson, “institutional image and retention,” tertiary education and man- agement , no. ( ): . . jeanie m. welch, “the electronic welcome mat: the academic library web site as a marketing and public-rela- tions tool,” the journal of academic librarianship , no. ( ): – . . oclc, “how academic librarians influence students’ web-based information choices,” in oclc online computer library center database online, ( ), , http://www .oclc .org/downloads/community/informationhabits.pdf (accessed march , ). . carnegie foundation, carnegie classification of institutions of higher education, edition, http://www.carnegiefound ation.org/classification/ (accessed jan. , ). . beth evans, “the authors of academic library home pages: their identity, training, and dissemination of web con- struction skills,” internet research , no. ( ): – . . oclc, . . u.s. department of justice, section home page, in united states department of justice database online, ( ), , http://www.usdoj.gov/crt/ / (accessed july , ). statement of ownership, management, and circulation information technology and libraries, publication no. - , is published quarterly in march, june, september, and december by the library information and technology association, american library association, e. huron st., chicago, illinois - . editor: john webb, librarian emeritus, washington state university libraries, pullman, wa - . annual subscription price, $ . printed in u.s.a. with periodical-class postage paid at chicago, illinois, and other locations. as a nonprofit organization authorized to mail at special rates (dmm section . only), the purpose, function, and nonprofit status for federal income tax purposes have not changed during the preceding twelve months. extent and nature of circulation (average figures denote the average number of copies printed each issue during the preceding twelve months; actual figures denote actual number of copies of single issue published nearest to filing date: june issue). total number of copies printed: average, , ; actual, , . sales through dealers and carriers, street vendors, and counter sales: none. paid or requested mail subscriptions: average, , ; actual, , . free distribution (total): average, ; actual, . total distribution: average, , ; actual, , . office use, leftover, unaccounted, spoiled after printing: average, ; actual, . total: average, , ; actual, , . percentage paid: average, . ; actual, . . s t a t e m e n t o f o w n e r s h i p , m a n a g e m e n t , a n d c i r c u l a t i o n ( p s f o r m , o c t o b e r ) f i l e d w i t h t h e u n i t e d s t a t e s p o s t o f f i c e p o s t m a s t e r i n c h i c a g o , s e p t e m b e r , . information technology and libraries | december l eadership—what is it? ala president leslie burger has me thinking about it a lot these days. as i write, the lita board is in the process of determining who lita will sponsor in the ala emerging leaders program. the task is difficult. lita has many new librarians who have strong potential for leadership. consequently i feel assured that lita has a strong future because what is an association, if not its members? so one of the questions the board asked was what does it mean to be an emerg- ing leader? when has one emerged? personally, i feel that i am still emerging because there is always more to learn. lifelong learning, isn’t that what librarians are all about? in preparation for my presidency, i attended an american society for association executives seminar facili- tated by tecker consultants. they defined four types of influential leadership: servant, visionary, expert, and cata- lytic. i see all four types of influential leaders within lita and they are all important. the servant leader provides service to others. in a volunteer organization like lita, a lot of servant leadership is being exhibited. these are the people who keep the organization humming, making sure we have the programs and education opportunities that make lita relevant to its members. the most obvi- ous place we see visionary leaders in lita is at our top technology trends; however, it is not the only place where visionary thinking occurs. lita members are often cutting- edge, applying new technologies to solve problems or to provide better solutions and services. visionary leadership is where one sees what the future could look like. lita programs are filled with expert leaders who share their technical expertise and lead the profession in applying those technologies. however, we also have many expert leaders who have important insights into what the asso- ciation can be. the catalytic leader brings people together and leverages their capabilities. the lita board works with other lita leadership to ensure that our goals are reached and to bring together all of the lita offerings to make membership a comprehensive professional benefit. my challenge as the current lita president with the ala emerging leaders program is to ensure that our sponsored member has a meaningful opportunity to become a superb leader both within lita and within the profession. in addition to attending the leadership training workshops for all of the emerging leaders, each sponsored person will be appointed to some service role within ala or one of its units. the lita board has elected to have our sponsored emerging leader work closely with the officers, in particular our vice president/president-elect mark beatty, on strategic planning for the next two years. i am hopeful that we will learn a great deal from our emerging leader regarding what new members are seeking out of the organization. when i think about a good leader, i think about some- one who listens, who allows others to think creatively and to take risks, who inspires, who sees the big picture, who can make decisions and make others understand the reasons for a decision, and who communicates well. john buchan put it this way: “the task of leadership is not to put greatness into people, but to elicit it, for the greatness is there already.” my goal this year, in conjunction with, but not limited to, the ala emerging leaders program, is to grow our new members into future lita leaders. i have been rewarded in all of my work within lita to wit- ness rising stars take on exciting roles and projects. i hope everyone reaps the joys of mentoring new professionals at some point in their careers. in my own leadership role, i take seriously the need to implement lita’s strategic plan. in that vein, the board has created an assessment and research task force that will make recommendations on gathering assessment data and feedback from members. with the appropriate knowl- edge base, we can ensure that value is being received. the board has also created a working group consisting of the chairs of the education committee, the regional institutes committee, and the program planning committee to make recommendations on our education programs. i have been working with that group to identify new modes of deliver- ing our programs and to ensure that they maintain their relevancy to lita members. lita continues to implement new communication technologies to reach out to its mem- bers. the lita blog has now been up for over a year and the new lita wiki is available for use by interest groups and others to allow experts to collaborate in the building of topic-specific resources. sir john harvey-jones framed the question thusly: “how do you know you have won? when the energy is coming the other way and when your people are visibly growing individually and as a group.” i see this happening in lita. what an energizing and fulfilling sight it is! president’s column bonnie postlethwaite bonnie postlethwaite (postlethwaiteb@umkc.edu) is lita pres- ident / and associate dean of libraries, university of missouri–kansas city. search across different media | buckland, chen, gey, and larson digital technology encourages the hope of searching across and between different media forms (text, sound, image, numeric data). topic searches are described in two different media: text files and socioeconomic numeric databases and also for transverse searching, whereby retrieved text is used to find topically related numeric data and vice versa. direct transverse searching across different media is impossible. descriptive metadata pro- vide enabling infrastructure, but usually require map- pings between different vocabularies and a search-term recommender system. statistical association techniques and natural-language processing can help. searches in socioeconomic numeric databases ordinarily require that place and time be specified. a hope for libraries is that new technology will support searching across an increasing range of resources in a growing digital landscape. the rise of the internet provides a technological basis for shared access to a very wide range of resources. the reality is that network-accessible resources, like the contents of a well-stocked reference library, are quite heterogeneous, especially in the variety of indexing, classification, catego- rization, and other forms of metadata. however, the use of digital technology implies a degree of technical compat- ibility between different media, sometimes referred to as “media convergence,” and these developments encourage the prospect of being able to search across and between different media forms—notably text, images, sound, and numeric data sets—for different kinds of material relat- ing to the same topic. to examine the practical problems involved, the authors undertook to demonstrate searching between and across two different media forms: text files and socioeconomic numeric data sets. two kinds of search are needed. first, it should be pos- sible to do a topical search in multiple media resources, so that one can find, for example, both pertinent factual numeric data and relevant discussion. (one difficulty is that the vocabulary used to classify the numeric data is ordinarily quite different from the subject headings used for books, magazine articles, and newspaper stories about the same topic.) second, when intriguing data values are encountered, one would like to move directly to topically relevant texts. likewise, when a questionable statement is read, one would like to be able to find relevant statisti- cal evidence. therefore, there needs to be search support that facilitates such transverse searching among resources, establishing connections, transferring data, and invoking appropriate utilities in a helpful way. both problems were addressed through the design and demonstration of a gateway providing search sup- port for both text and socioeconomic numeric databases. first, the gateway should help users conduct searches in databases of different media forms by accepting a query in the searcher’s own terms and then suggesting the spe- cialized categorization terms to search for in the selected resource. second, if something interesting was found in a socioeconomic database, the gateway would help the searcher to find documents on the same topic in a text database, and vice versa. selection of the best search terms in target databases is supported by the use of indexes to the categories (entries, headings, class numbers) in the system to be searched. these search-term recommender systems (also known as “entry vocabulary indexes”) resemble dewey’s “relativ index,” but are created using statistical association techniques. four characteristics of this investigation need to be noted: . searching independent sources: the authors were not concerned with ingesting resources from differ- ent sources into a consolidated local data repository and searching within it. the interest lay, instead, in being able to search effectively in any accessible resource as and when one wants. this implies that interoperability issues in dealing with the native query languages and metadata vocabularies of remote repositories can be solved. . search for independent content: numeric data sets commonly have associated text in the form of documentation, code books, and commentary. however, the authors were interested in finding topical content that had no such formal or liter- ary connection. independent means, for example, a newspaper article written by someone unaware that relevant statistical data existed or had been written before the author’s article existed. in the other direction, having found statistical data of interest, could topically related text created inde- pendently of this particular data point be found? . two different media forms were chosen: text and numeric data sets. they look similar because they both use arabic numerals, but the traditional reli- ance on information retrieval in a text environment search across different media: numeric data sets and text files michael buckland, aitao chen, fredric c. gey, and ray r. larson michael buckland (buckland@sims.berkeley.edu) is emeritus professor, school of information, university of california, berkeley; aitao chen (aitao@yahoo-inc.com) is a researcher at yahoo!, sunnyvale, california; fredric c. gey (gey@berkeley .edu) is an information scientist, uc data archive and technical assistance at the university of california, berkeley; and ray r. larson (ray@sims.berkeley.edu) is a professor, school of information at the university of california, berkeley. information technology and libraries | december of using any character string from the corpus as a query, although technically feasible, cannot be expected to be useful here. one can copy a number expressing quantity, such as , , from a numeric data cell, use it as a query in a text search engine such as google, and retrieve a large and eclectic retrieved set, usually involving “ ” as an iden- tifying number for a postal code, a memorandum, a part number, software bug report, and so on, but the relationship is spurious. it requires great faith in numerology to expect anything topically mean- ingful to the original data cell one started with. with other combinations of media forms, not even spurious results are feasible: one cannot submit a musical fragment or some pixels from an image as a text query. . the authors’ interest was in how to achieve a bet- ter return on existing investments in well-formed, edited resources with descriptive metadata. this project built directly on prior work on how to make more effective use of existing, expertly developed metadata, rather than creating or replacing meta- data. search of multiple resources comes in two forms: . parallel search is when a single query is sent to two or more resources at more or less the same time. for example, a researcher interested in the import of shrimp would like to see pertinent newspaper articles and trade statistics. thus, one might send a query to the census bureau’s united states (u.s.) imports and exports numeric data series and look at sic for shrimp and prawn and note a dra- matic increase in imports from vietnam through los angeles from onwards. one would also search newspaper indexes for articles such as “normalizing ties to vietnam important steps for u.s. firms; california stands to profit handsomely when barriers fall to trade with fast-growing coun- try.” different sources are likely to use different index terms or categories, so the challenge is how to express the searcher’s query in terms that will be effective for searching in the target resources, which, mostly likely, will use different vocabular- ies. as one example, the term for “automobiles” is in the standard industrial classification; tl in the library of congress (lc) classification, / in the u.s. patent classification; and, in the census bureau’s u.s. imports and exports data series, pass mot veh, spark ign eng. . transverse search is when an item of interest found in one resource is used as the basis for a query to be forwarded to a different resource. the challenge here, again, is that when a query using the topical metadata in one resource needs to be expressed in the vocabulary of the target resource, the metadata vocabularies in the two resources will usually be different from each other, and, quite likely, both are unfamiliar to the searcher. when searching within a single media form, it may be possible to use content itself directly as a query: a frag- ment of text in a source-text database is commonly used as a query in a target-text database. similarly, one might start with an image and seek images that are measur- ably similar. however, because such direct search cannot be done when searching across different media forms, an indirect approach relying on the use of interpretive representations becomes necessary. as the network envi- ronment expands, mapping between vocabularies will be increasingly important. ■ text and numeric resources text resource a library catalog—a special case of text file—was chosen for use as a text file rather than a corpus of “full text.” the reasons were practical: in this exploratory investiga- tion, it was important to start with resources that had rich metadata; it needed to be a resource that was sufficiently controllable to enable experimentation with it. a library catalog was in the spirit of the project in that it would lead to additional text resources; and a suitable resource was available, which was intended for metadata mapping: a set of several million marc records, derived from melvyl, the university of california online library catalog. socioeconomic numeric data set initially, and in prior work, the authors had worked on access to u.s. federal data series, especially import and export statistics and county business reports. although some progress was made with interfaces to these data series, it became clear that the investment needed to craft interoperable access was high relative to the available staff. crafting access to individual data series did not appear to be a scalable way to demonstrate variety within the authors’ limited resources, so attention was turned to a single collection comprising many diverse numeric tables, the counting california database. ■ mapping topical metadata well-edited, high-quality databases typically have topi- cal metadata expertly assigned from a vocabulary (the- saurus, classification, subject-heading system, or set of search across different media | buckland, chen, gey, and larson categories). but there is a babel of different vocabularies. not only do the names of topics vary, but the underlying concepts or categories may also differ. effective searching requires expert familiarity with a system’s vocabulary; but as access to digital resources expands, the diversity of vocabularies increases and accessible resources are decreasingly likely to use vocabularies familiar to any individual searcher. the best answer is twofold: first, it is desirable to have an index (a “mapping”) from the natural language of each group of searchers to the entries used in each metadata vocabulary. such a mapping provides an index from a vocabulary familiar to the searcher to the vocabulary used in entries of the target system and so is called a search-term recommender system. (the authors called it an “entry-vocabulary index,” or evi.) dewey’s “relativ index” to his decimal classification is a famil- iar example. when searching across databases, one also wants a second kind of mapping: between pairs of system vocabularies. unfortunately, mappings between different vocabularies are rare, expensive, time-consuming, and hard to maintain. (the unified medical language system is a notable example.) it is the authors’ impression that this problem is worse in searching across different media forms because data bases in different media forms tend to be created by different communities, increasing the chances that they will use different categories, vocabularies, and ways of thinking. fortunately where data containing two forms of vocabulary are available, they can be used as training sets for statistical-association techniques to generate evis auto- matically, and this is the approach that was used. (more details can be found in the appendix.) from text words to library subject headings an evi from ordinary english words to library of congress subject headings (lcsh) was created by taking catalog records containing at least one subject heading ( xx field in the marc bibliographic format). from each of the , , records used, main subject headings were extracted (subfield a from fields , , , , , and ) and fields containing text: titles ( a), subtitles ( b), and summaries describing the scope and general content of the material ( a). the underlying assump- tion is that for each record, the words in the “text” fields ( a,b and a) tend to be characteristic of discourse on the subject ( xxa). two examples, with identifying lccns in the < > field are: < > //r < >a study of operant conditioning under delayed reinforcement in early infancy < >infant psychology < >operant conditioning < > < >reptilian diseaserecognition and treatment < >reptilesdiseases the words in the text fields ( a, b, and a) were extracted. stop words were removed and the remainder normalized. then the degree to which each word is asso- ciated with each subject heading (by co-occurring in the same records) was computed using a maximum likelihood ratio-based measure. natural-language processing can be used to identify adjective-noun phrases to support more precise searching using phrases as well as individual words. a very large matrix shows the association of each text word (or phrase) with each subject heading; so, for any given word (or combination of words), a list of the most closely associated headings, ranked by degree of association, can be derived from the matrix. queries a query, which can be a single word, a phrase, a set of keywords, a book title, and so on, is normalized in the same way and looked up in the matrix to produce a ranked list of the most closely associated subject headings as candidate lcsh search terms. for example, entering the textual query words “peanut” and “butter” generates the following ranking list of lcsh main headings as candi- dates for searching: rank lcsh (subfield a) . peanut . cookery (peanut butter) . cookery (peanuts) . peanut industry . peanut butter . butter . schulz, charles m. this display is an important departure from traditional fully automatic searching. the list is, in effect, a prompt, indicating probably suitable query terms in the vocabulary of the target resource. it introduces the searcher to the categories and terminology of the system and enables the searcher to use expert judgment to select the heading that seems best for the search. from text words to the metadata vocabularies in numeric data sets a training set of records containing both descriptive words and topical metadata is often not readily available for numeric data sets. the authors’ first effort was to create an evi to the standard industrial classification (sic), widely used over many years in numeric data sets. (sic codes were associated with words by using, as a training set, the information technology and libraries | december titles in a bibliographic database that used sic codes.) but by the time the sic evi was completed, sic had been dis- continued and replaced by the north american industry classification system (naics), so a mapping was created from sic codes to naics codes. figures – show stages in an interface that accepts a searcher’s query “car” (figure ), prompts with a ranked list of naics codes (figure ), then extends the search with the selected naics code to retrieve numeric data (figure ). by this time, however, it had become apparent that, with the current low level of interoperability in software and in data formats, the labor required to create evis and interfaces to each large traditional numeric data series was enormous. therefore, attention was turned to a collection of different numeric data sets available through a single interface, counting california, made available by california digital library at http://countingcalifornia.cdlib.org. this resource is a collection of some three thousand numeric tables containing statistics related to a range of topics. the numeric data sets are mainly from the california department of health services, the california department of finance, and the federal bureau of the census. the tables are organized under a two-level classification scheme. there are sixteen topics at the top level, which are subdi- vided into a total of subtopics. all the numeric tables were assigned to one or more subtopics and each table has a caption. at the counting california web site, a searcher can browse for tables by selecting a higher-level topic, then a lower-level subtopic, and then a table. two additional ways were created to access the tables: probabilistic retrieval, and an evi to the topical categories. the cap- tions, topics, and subtopics were extracted for each of the three thousand tables, and xml records were created in the following form: education libraries
library statistics, statewide summary by type of library california – to –
retrieval two search methods were used: direct probabilistic retrieval. an in-house implementa- tion was used of a probabilistic full-text retrieval algo- rithm developed at berkeley. this search engine takes a free-form text query and returns a ranked list of captions of tables ranked according to their relevance scores. for example, the five top-ranked captions returned to the query “public libraries in california” were: figure . query interface for search-term recommender system f or the north american industry classification system figure . display of naics code search-term recommendations for “car” figure . display of numeric data retrieved using selected naics code search across different media | buckland, chen, gey, and larson . library statistics, statewide summary by type of library california, – to – table f . . library statistics, statewide summary by type of library california, – to – table f yr - . . number of california libraries, to table f yr . number of california libraries, to , as of september table f . . california public schools, grades k– , to table f . each entry in the retrieved set list is linked to a numeric table maintained at the counting california web site and, by clicking on the appropriate link, a user can display the table as an ms excel file or as a pdf file. mediated search. from the same extracted records the words in the captions were used to create an evi to the sub- topics in the topic classification using the method already described. as an example, the query “personal individual income tax,” when submitted to the evi, generated the following ranked list of subtopics: . income . government earnings and tax revenues . personal income . property tax . personal income tax . corporate income tax . per capita income a user can click on any selected subtopic to retrieve the cap- tions of tables assigned that subtopic. for example, clicking on the fifth subtopic, personal income tax, retrieves: ■ personal income tax returns: number and amount of adjusted gross income reported by adjusted gross income class california, taxable year. table d yr ■ personal income tax returns: number and amount of adjusted gross income reported by adjusted gross income class california, taxable year. table d ■ personal income statistics by county, california taxable year. table d ■ personal income statistics by county, california taxable year. table d yr ■ transverse searching between text- and numeric-data series to demonstrate the searching capability from a bib- liographic record to numeric-data sets, the first step is to retrieve and display a bibliographic record from an online catalog. a web-based interface for searching online catalogs was implemented using an in-house implementation of the z . protocol. besides the z . protocol, an important component that makes searching remote online catalogs feasible is the gateway between the http (hypertext transfer protocol) and the z . protocol. while http is a connectionless-oriented protocol, the z . is a connec- tion-oriented protocol. the gateway maintains connections to remote z . servers. all search requests to any remote z . server go through the gateway. searching from catalog records to numeric data sets having selected some text (for the purposes of this study, a catalog record), how could one identify the facts or statis- tics in a numeric database that are most closely related to the topic? clicking on a “formulate query” button placed at the end of a displayed full marc record creates a query for searching a numeric database. the initial query will contain the words extracted from the title, subtitle, and the subject headings and is placed in a new window where the user can modify or expand the query before submitting it to the search engine for a numeric database. so, for example, the following text extracted from a catalog record: library laws of the state of california, library legislation. california. public libraries when submitted as a query, retrieves a ranked list of table names, of which two, covering different time periods, are entitled library statistics, statewide summary by type of library, california. searching from numeric data sets from catalog records transverse search in the other direction, starting from a data table, is achieved by forwarding the caption of a table to the word-to-lcsh evi to generate a prompt list of the seven top-ranked lchss, any one of which can be used as a query submitted to the catalog. ■ architecture figure shows the structure of the implementation. the boxes shown in the figure are: . a search interface for accessing bibliographic/tex- tual resources through a word-to-lcsh evi. . a word to the lcsh evi. . a ranked list of lcshs closely associated with the query. . an online catalog. information technology and libraries | december . results of searching the online catalog using an lcsh. . a full marc record displayed in tagged form. . a new query formed by extracting the title and sub- ject fields from the displayed full marc record. . a numeric database. . a list of captions of numeric tables ranked by rel- evance score to the query. . numeric table displayed in pdf or ms excel for- mat. . a search interface for numeric databases based on a probabilistic search algorithm. a user can start a search using either interface (boxes or ) and, from either starting point, find records on the same topic of interest in a textual (here bibliographic) database and a socioeconomic database. ■ conclusions and further work enhanced access to numeric data sets the descriptive texts associated with numeric tables, such as the caption, headers, or row labels, are usually very short. they provide a rather limited basis for locating the table in response to queries, or describing a data cell sufficiently to form a usefully descriptive query from it. sometimes the title (caption) of a table may be the only searchable textual description about the content of the table, and the titles are sometimes very general. for example, one of the titles, library statistics, statewide summary by type of library california, – to – , is so general that neither the kinds of statistics nor the types of libraries are revealed. if a user posed the question, “what are the total operating expenditures of public libraries in california?” to a query system that indexes table titles only, the search may well be ineffective since the only word in common between the table title and the user’s query is “california” and, if the plurals of nouns have been normalized, to the singular form, “library.” table column headings and row headings provide additional information about the content of a numeric table. however, the column and row headings are usu- ally not directly searchable. for example, a table named “language spoken at home” in counting california databases consists of rows and columns. the column headings list the languages spoken at home, while the row headings show the county names in california. each cell in the table gives the number of people, five years of age and older, who speak a specific language at home. to answer questions such as “how many people speak spanish at home in alameda county, california?” using the table title alone may not retrieve the table that contains the answer to the example question. it is recommended that the textual descriptions of numeric tables be enriched. automatically combining the table title and its column and row headings would be a small but practical step toward improved retrieval. geographic search socioeconomic numeric data series refer to particular areas and, in contrast to text searching, the geographical aspect ordinarily has to be specified. to match the geographical area of the numeric data, a matching text search may also have to specify the same place. the authors found that this was hard to achieve for several reasons. place names are ambiguous and unstable: a search for data relating to trinidad might lead to trinidad, west indies, instead of trinidad, california, for example. the problem is compounded because, in numeric data series, specialized geopolitical divisions, such as census tracts and counties, are commonly used. these divisions do not match conve- niently with searchers’ ordinary use of place names. also, the granularity of geographical coverage may not match well. data relating to berkeley, for example, may be avail- able only in aggregated data for alameda county. it was eventually concluded that reliance on the names of places could never work satisfactorily. the only effective path to reliable access to data relating to places would be to use geospatial coordinates (latitude and longitude) to establish unambiguously the identity and location of any place and the relationship between places. this means that gazetteers and map visualizations become important. gazetteers relate named places to defined spaces, and thereby reveal spatial relationships between places, e.g., the city of alameda is on alameda island within alameda county. this problem has been addressed in a subsequent figure . architecture of the prototype search across different media | buckland, chen, gey, and larson study entitled “going places in the catalog: improved geographical access.” temporal search searches of text files and of socioeconomic numeric data series also differ substantially with respect to time periods: numeric data searches ordinarily require the years of inter- est to be specified; text searches rarely specify the period. an additional difficulty arises because in text, as in speech, a period is commonly referred to by a name derived meta- phorically from events used as temporal markers, rather than by calendar time, as in “during vietnam,” “under clinton,” or “in the reign of henry viii.” named time periods have some of the characteristics of place names: they are culturally based and tend to be multiple, unstable, and ambiguous. it appears that an analogous solution is indicated: directories of named time periods mapped to calendar definitions, much as a gazet- teer links place names to spatial locators. this problem is being addressed in a subsequent study entitled “support for the learner: what, where, when, and who.” media forms the paradox, in an environment of digital “media conver- gence,” that it appears impossible to search directly across different media forms invites closer attention to concepts and terminology associated with media. a view that fits and explains the phenomena as the authors understand them, distinguishes three aspects of media: ■ cultural codes: all forms of expression depend on some shared understandings, on language in a broad sense. convergence here means cultural convergence or interpretation. ■ media types: different types of expression have evolved: texts, images, numbers, diagrams, art. an initial classification can well start with the five senses of sight, smell, hearing, taste, and feel. ■ physical media: paper; film; analog magnetic tape; bits; . . . being digital affects directly only this aspect. anything perceived as a meaningful document has cul- tural, type, and physical aspects, and genre usefully denotes specific combinations of code, type, and physical medium adopted by social convention. genres are historically and culturally situated. convergence can be understood in terms of interoper- ability and is clearly seen in physical media technology. the adoption of english as a language for international use in an increasingly global community promotes conver- gence in cultural codes. nevertheless, the different media types are fundamentally distinct. metadata as infrastructure it is the metadata and, in a very broad sense, “biblio- graphic” tools that provide the infrastructure necessary for searches across and between different media—thesauruses, mappings between vocabularies, place-name gazetteers, and the like. in isolation, metadata is properly regarded as description attached to documents, but this is too narrow a view. collectively, the metadata forms the infrastructure through which different documents can be related to each other. it is a variation on the role of citations: individually, references amplify an individual document by validating statements made within it; collectively, as a citation index, references show the structure of scholarship to which docu- ments are attached. ■ summary a project was undertaken to demonstrate simultane- ous search of two different media types (socioeconomic numeric data series and text files) without ingesting these diverse resources into a shared environment. the project objective was eventually achieved, but proved harder than expected for the following reasons: access to these differ- ent media types has been developed by different commu- nities with different practices; the systems (vocabularies) for topical categorization vary greatly and need interpre- tative mappings (also known as relative indexes, search- term recommender systems, and evis); specification of geographical area and time period are as necessary for search in socioeconomic data series and, for this, existing procedures for searching text files are inadequate. ■ acknowledgement this work was partially supported by the institute of museum and library services through national library leadership grant no. for a project entitled “seamless searching of numeric and textual resources,” and was based on prior research partially supported by darpa contracts n - -c- ; ao# f : “search support for unfamiliar metadata vocabularies” and n - - - , to# j : “translingual information management using domain ontologies.” references . michael k. buckland, fredric c. gey, and ray r. larson, seamless searching of numeric and textual resources: final report on institute of museum and library services national leadership information technology and libraries | december grant no. (berkeley, calif.: univ. of california, school of information management and systems, ), http:// metadata.sims.berkeley.edu/papers/seamlesssearchfinal report.pdf (accessed july , ); michael buckland et al., “seamless searching of numeric and textual resources: fri- day afternoon seminar, feb. , ,” http://metadata.sims .berkeley.edu/papers/seamlessfri.ppt (accessed july , ). . michael buckland et al., “mapping entry vocabulary to unfamiliar metadata vocabularies,” d-lib magazine , no. (jan. ), www.dlib.org/dlib/january /buckland/ buckland .html (accessed july , ); michael buckland, “the sig- nificance of vocabulary,” , http://metadata.sims.berkeley .edu/vocabsig.ppt (accessed july , ); fredric c. gey et al., “entry vocabulary: a technology to enhance digital search,” in proceedings of the first international conference on human lan- guage technology, san diego, mar. (san francisco: morgan kaufmann, ), – , http://metadata.sims.berkeley.edu/ papers/hlt -final.pdf (accessed july , ). . los angeles times, july , : d . . michael buckland, “vocabulary as a central concept in library and information science,” in digital libraries: interdisci- plinary concepts, challenges, and opportunities. proceedings of the third international conference on conceptions of library and infor- mation science (colis ), dubrovnik, croatia, may – , , ed. t. arpanac et al. (lokve, croatia: benja pubs., ), – , www .sims.berkeley.edu/~buckland/colisvoc.htm (accessed july , ); buckland et al., “mapping entry vocabulary.” . counting california, http://countingcalifornia.cdlib.org (accessed july , ). . “factsheet: unified medical language system,” www .nlm.nih.gov/pubs/factsheets/umls.html (accessed july , ). . william s. cooper, aitao chen, and fredric c. gey, “full- text retrieval based on probabilistic equations with coefficients fitted by logistic regression,” in d. k. harman, ed., the second text retrieval conference (trec- ), march , – (gaith- ersburg, md.: national institute of standards and technol- ogy, ), http://trec.nist.gov/pubs/trec /papers/txt/ .txt (accessed july , ). . “going places in the catalog: improved geographical access,” http://ecai.org/imls (accessed jul. , ). . vivien petras, ray larson, and michael buckland, “time period directories: a metadata infrastructure for placing events in temporal and geographic context,” in opening information horizons: joint conference on digital libraries (jcdl), chapel hill, n.c., june – , , forthcoming, http://metadata.sims .berkeley.edu/tpdjcdl .pdf (accessed july , ); “support for the learner: what, where, when, and who,” http://ecai .org/imls (accessed july , ). search across different media | buckland, chen, gey, and larson appendix: statistical association methodology a statistical maximum likelihood ratio weighting tech- nique was used to construct a two-way contingency table relating each natural-language term (word or phrase) with each value in the metadata vocabulary of a resource, e.g., lcsh, lccns, u.s. patent classification numbers, and so on. an associative dictionary that will map words in natural languages into metadata terms can also, in reverse, return words in natural language that are closely associated with a metadata value. training records containing two different metadata vocabularies can be used to create direct mappings between the values of the two metadata vocabularies. for example, u.s. patents contain both u.s. and international patent classification numbers and so can be used to create a mapping between these two quite different classifica- tions. multilingual training sets, such as catalog records for multilingual library collections, can be used to create multilingual natural language indexes to metadata vocabu- laries and, also, mappings between natural language vocabularies. in addition to the maximum likelihood ratio-based association measure, there are a number of other asso- ciation measures, such as the chi-square statistic, mutual information measure, and so on, that can be used in creat- ing association dictionaries. the training set used to create the word-to-lcsh evi was a set of catalog records with at least one assigned lcsh (i.e., at least one xx field). natural language terms were extracted from the title (field a), subtitle ( b), and summary note ( a). these terms were tokenized; the stopwords were removed; and the remaining words were normalized. a token here can contain only letters and digits. all tokens were then changed to lower case. the stoplist has about six hundred words considered not to be content bearing, such as pronouns, prepositions, coordinators, determiners, and the like. the content words (those not treated as stopwords) were normalized using a table derived from an english morphological analyzer. the table maps plural nouns into singular ones; verbs into the infinitive form; and comparative and superlative adjectives to the positive form. for example, the plural noun printers is reduced to printer, and children to child; the comparative adjective longer and the superlative adjective longest are reduced to long; and printing, printed, and prints are all reduced to the same base form print. when a word belonging to more than one part-of-speech category can be reduced to more than one form, it is changed to the first form listed in the morphological analyzer table. as an example, the word saw, which can be a noun or the past tense of the verb to see, is not reduced to see. subject headings (field xxa) were extracted without qualifying subdivisions. the inclusion of foreign words (alcoholismo, alcoolisme, alkohol, and alcool), derived from titles in foreign languages, demonstrate that the technique is language independent and could be adopted in any country. it could also support diversity in u.s. libraries by allowing searches in spanish or other languages, so long as the training set contains sufficient content words. evis are accessible at http://metadata. sims.berkeley.edu/prototypesi.html. fuller descriptions of the project methodology can be found in the literature. ■ references . ted dunning, “accurate methods for the statistics of surprise and coincidence,” computational linguistics (march ): – . . daniel karp et al., “a freely available wide cover- age morphological analyzer for english,” in proceedings of coling- , nantes, (morristown, n.j.: association for computational linguistics, ), – , http://acl.ldc.upenn .edu/c/c /c - .pdf (accessed july , ). . michael k. buckland, fredric c. gey, and ray r. larson, seamless searching of numeric and textual resources: final report on institute of museum and library services national leadership grant no. (berkeley, calif.: univ. of california, school of informa- tion management and systems, ), http://metadata.sims .berkeley.edu/papers/seamlesssearchfinalreport.pdf (accessed jul. , ); youngin kim et al., “using ordinary language to access metadata of diverse types of information resources: trade classification and numeric data,” in knowledge: creation, organization, and use. proceedings of the american society for infor- mation science annual meeting, oct. –nov. , (medford, n.j.: information today, ), – . levan opensearch and sru | levan not all library content can be exposed as html pages for harvesting by search engines such as google and yahoo!. if a library instead exposes its content through a local search interface, that content can then be found by users of metasearch engines such as a and vivísimo. the functionality provided by the local search engine will affect the functionality of the metasearch engine and the findability of the library’s content. this paper describes that situation and some emerging standards in the metasearch arena that choose different balance points between functionality and ease of implementation. editor's note: this article was submitted in honor of the fortieth anniversaries of lita and ital. ฀ the content provider’s dilemma consider the increasingly common situation in which a library wants to expose its digital content to its users. sup- pose it knows that its users prefer search engines that search the contents of many sites simultaneously, rather than site-specific engines such as the one on the library’s web site. in order to support the preferences of its users, this library must make its contents accessible to search engines of the first type. the easiest way to do this is for the library to convert its contents to html pages and let the harvesting search engines such as google and yahoo! collect those pages and provide searching on them. however, a serious problem with harvesting search engines is that they place limits on how much data they will collect from any one site. google and yahoo! will not harvest a -million-record book catalog, even if the library can figure out how to turn the catalog entries into individual web pages. an alternative to exposing library content to harvest- ing search engines as html pages is to provide a local search interface and let a metasearch engine combine the results of searching the library’s site with the results from searching many other sites simultaneously. users of metasearch engines get the same advantage that users of harvesting search engines get (i.e., the ability to search the contents of many sites simultaneously) plus those users get access to data that the harvesting search engines do not have. the issue for the library is determining how much functionality it must provide in its local search engine so that the metasearch engine can, in turn, provide accept- able functionality to its users. the amount of functionality that the library provides will determine which metasearch engines will be able to access the library’s content. metasearch engines, such as a and vivísimo, are search engines that take a user’s query, send it to other search engines, and integrate the responses. the level of integration usually depends on the metasearch engine’s ability to understand the responses it receives from the various search engines it has queried. if the response is html intended for display on a browser, then the metasearch engine developers have to write code to parse through the html looking for the content. in such a case, the perceived value of the content determines the level of effort that the metasearch engine developers put into the parsing task; low-value content will have a low priority for developer time and will either suffer from poor integration or be excluded. for metasearch engines to work, they need to know how to send a search to the local search engine and how to interpret the results. metasearch engines such as vivísimo and a have staffs of programmers who write code to translate the queries they get from users into queries that the local search engines can accept. metasearch engines also have to develop code to convert all the responses returned by the local search engines into some common format so that those results can be combined and displayed to the user. this is tedious work that is prone to breaking when a local search engine changes how it searches or how it returns its response. the job of the metasearch engine is made much simpler if the local search engine supports a standard search interface such as sru (search and retrieve url) or opensearch. ฀ what does a metasearch engine need in order to use a local search engine? the search process consists of two basic steps. first, the search is performed. second, records are retrieved. to do a search, the metasearch engine needs to know: . the location of the local search engine . the form of the queries that the local search engine expects . how to send the query to the local search engine to retrieve records, the metasearch engine needs to know: . how to find the records in the response . how to parse the records opensearch and sru: a continuum of searching ralph levan ralph levan (levan@oclc.org) is a research scientist at oclc online computer library center in dublin, ohio. information technology and libraries | september ฀ four protocols this paper will discuss four search protocols: opensearch, opensearch . , sru, and the metasearch xml gateway (mxg). opensearch was initially developed for the a meta- search engine. it provides a mechanism for content providers to notify a of their content. it also allows rss (really simple syndication) browsers to display the results of a search. opensearch . has just been released. it extends the original specification based on input from a number of organizations, microsoft being prominent among them. sru was developed by the z . community. recognizing that their standard (now eighteen years old) needed updating, they simplified it and created a new web service based on an xml encoding carried over http. the mxg protocol is the product of the niso metasearch initiative, a committee of metasearch engine developers, content providers, and users. mxg uses sru as a starting place, but eases the requirement for support of a standard query grammar. ฀ functionality versus ease of implementation a library rarely has software developers. the library’s area of expertise is, first of all, the management of content and, sec- ondarily, content creation. librarians use tools developed by other organizations to provide access to their content. these tools include the library’s opac, the software provided to search any licensed content, and the software necessary to build, maintain, and access local digital repositories. for a library, ease of adoption of a new search protocol is essential. if support for the search protocol is built into the library’s tools, then the library will use it. if a small piece of code can be written to convert the library’s existing tools to support the new protocol, the library may do that. similarly, the developers of the library’s tools will want to expend the minimum effort to support a new search protocol. the tool developer’s choice of search protocol to sup- port will depend on the tension between the functionality needed and the level of effort that must be expended to provide and maintain it. if low functionality is acceptable, then a small development effort may be acceptable. high functionality will require a greater level of effort. the developers of the search protocols examined here recognize this tension and are modifying their protocols to make them easier to implement. the new opensearch . will make it easier for some local search-engine providers to implement by easing some of the functionality require- ments of version . . similarly, the niso metasearch committee has defined mxg, a variant of sru that eases some of the requirements of sru. ฀ search protocol basics once again, the five basic pieces of information that a metasearch engine needs in order to communicate effec- tively with a local search engine are: ( ) local search engine location, ( ) the query-grammar expected, ( ) the request encoding, ( ) the response encoding, and ( ) the record encoding. the four protocols provide these pieces of infor- mation to one degree or another (see table ). the four protocols expose a site’s searching functional- ity and return responses in a standard format. all of these protocols have some common properties. they expect that the content provider will have a description record that describes the search service. all of these services send searches via http as simple urls, and the responses are sent back as structured xml. to ease implementation, opensearch . allows the content provider to return html instead of xml. all four protocols use a description record to describe the local search engine. the opensearch protocols define what a description record looks like, but not how it is retrieved. the location of the description record is dis- covered by some means outside the protocol (a priori knowledge). the description record specifies the location of the local search engine. the sru protocols define what a description record looks like and specifies that it can be obtained from the local search engine. the location of the local search engine is provided by a means outside the protocol (a priori knowledge again). each protocol defines how to formulate the search url. opensearch does this by having the local search-engine provider supply a template of the url in the description record. sru does this by defining the url. opensearch and mxg do not define how to formu- late the query. the metasearch engine can either pass the user’s query along to the local search engine unchanged or reformulate the query based on information about the local search engine’s query language that it has gotten by outside means (more a priori knowledge). in the first case, the metasearch engine has to hope that some magic will happen and the local search engine will do something useful with the query. in the latter case, the metasearch engine’s staff has to develop a query translator. sru specifies a standard query grammar: cql (common query language). this means that the meta- search engine only has to write one translator for all the sru local search engines in the world. but it also means that all the sru local search engines have to support the cql query grammar. since there are no local search engines that support cql as their native query grammar, the con- tent provider is left with the task of translating cql que- ries into their native query grammar. the query translation task has moved from the metasearch engine to the content provider. opensearch and sru | levan opensearch . , mxg, and sru define the struc- ture of the query response. in the case of opensearch, the response is returned as an rss message, with a couple of extra elements added. mxg and sru define an xml schema for their responses. opensearch . allows the local search engine to return the response as unstructured html. this moves the requirement of creating a standard response from the content provider and leaves the metasearch engine with the much tougher task of finding the content embedded in html. if the metasearch engine doesn’t write code to parse the response, then all it can do is display the response. it will not be able to combine the response from the local search engine with the responses from other engines. sru and mxg require that records be returned in xml and that the local search engine must specify the schema for those records in the response. this leaves the content provider with the task of formatting the records according to the schema of their choice, a task that the content provider is probably best able to do. in turn, the metasearch engine can convert the returned records into some common format so that the records from multiple local search engines can be combined into a single response. because the records are encoded in xml, it is assumed that standard xml format- ting tools can be used for the conversion. opensearch does not define how records should be structured. the opensearch response has a place for the title of the record and a url that points to the record. the structure of the record is undefined. this leaves the metasearch engine with the task of parsing the record that is returned. again, the effort moves from the content pro- vider to the metasearch engine. if the metasearch engine does not or cannot parse the records, then it can at least display the records in some context, but it cannot combine them with the records from another local search engine. ฀ conclusion these protocols sit on a spectrum of complexity, trading the content provider’s complexity for that of the search engine. however, with lessened complexity for the metasearch engine comes increased functionality for the user. metasearch engines have to choose what content providers they will search. those that provide a high level of functionality can be easily combined with their existing local search engines. content providers with a lower level of functionality will either need additional development by the metasearch engine or will not be searched. not all metasearch engines require the same level of functionality, nor will they be prepared to accept content with a low level of functionality. content providers, such as digital librar- ies and institutional repositories, will have to choose the functionality they need to support to reach the metasearch engines they desire. references and notes . joe barker, “meta-search engines,” in finding information on the internet: a tutorial (u.c. berkeley: teaching library inter- net workshops, aug. , [last update]), www.lib.berkeley. edu/teachinglib/guides/internet/metasearch.html (accessed may , ). . a .com, “opensearch specification,” http://opensearch .a .com/spec/ (accessed may , ); a .com, “opensearch . ,” http://opensearch.a .com/spec/ . / (accessed may , ). . mark pilgrim, “what is rss?” o’reilly xml.com, dec. , , www.xml.com/pub/a/ / / /dive-into-xml.html (accessed may , ). . the library of congress network development and marc standards office, “z . maintenance agency page,” www.loc.gov/z /agency/ (accessed may , ). . national information standards organization, “niso metasearch initiative,” www.niso.org/committees/ ms_initiative.html (accessed may , ). . niso metasearch initiative task group , “niso metase- arch xml gateway implementors guide, version . ,” may , , [microsoft word document] www.lib.ncsu.edu/niso- mi/images/ / /niso_metasearch_initiative_xml _gateway _implementors_guide.doc (accessed may , ); the library of congress, “sru: search and retrieve via url; sru version . february ,” www.loc.gov/standards/sru/index.html (accessed may , ). . the library of congress, “common query language; cql version . th february .” [web page] www.loc .gov/standards/sru/cql/index.html (accessed may , ). table . comparison of requirements of four metasearch protocols for effective communication with local search engines protocol feature opensearch . opensearch . mxg sru local search engine location a priori a priori a priori a priori request encoding defined defined defined defined response encoding none rss xml xml record encoding none none xml xml query grammar none none none cql manzari user-centered design of a web site | manzari and trinidad-christensen this study describes the life cycle of a library web site created with a user-centered design process to serve a graduate school of library and information science (lis). findings based on a heuristic evaluation and usability study were applied in an iterative redesign of the site to better serve the needs of this special academic library population. recommendations for design of web-based services for library patrons from lis programs are dis- cussed, as well as implications for web sites for special libraries within larger academic library settings. u ser-centered design principles were applied to the creation of a web site for the library and information science (lis) library at the c. w. post campus of long island university. this web site was designed for use by master’s degree and doctoral students in the palmer school of library and information science. the prototype was subjected to a usability study consisting of a heuristic evaluation and usability testing. the results were employed in an iterative redesign of the web site to better accommodate users’ needs. this was the first usabil- ity study of a web site at the c. w. post library. human-computer interaction, the study of the inter- action of human performance with computers, imposes a rigorous methodology on the process of user-interface design. more than an intuitive determination of user- friendliness, a successful interactive product is developed by careful design, testing, and redesign based on the testing outcomes. testing the product several times as it is being developed, or iterative testing, allows the users’ needs to be incorporated into the design. the interface should be designed for a specific community of users and set of tasks to be accomplished, with the goal of creating a consistent, usable product. the lis library had a web site that was simply a description of the collection and did not provide access to online specialized resources. a new web site was designed for the lis library by the incoming lis librarian who made a determination of what content might be useful for lis students and faculty. the goal was to have such content readily accessible in a web site separate from the main library web site. the web site for the lis library includes: ฀ access to all online databases and journals related to lis; ฀ a general overview of the lis library and its resources as well as contact information, hours, and staff; ฀ a list of all print and online lis library journal sub- scriptions, grouped by both title and subject, with links to access the online journals; ฀ links to other web sites in the lis field; ฀ links to other university web pages, including the main library’s home page, library catalog, and in- structions for remote database access, as well as to the lis school web site; ฀ a link to jake (jointly administered knowledge environment), a project by yale university that allows users to search for periodical titles within online databases, since the library did not have this type of access through its own software. this information was arranged in four top-level pages with sublevels. design considerations included making the site both easy to learn and efficient once users were familiar with it. since classes are taught at four locations in the metropolitan area, the site needed to be flexible enough to serve students at the c. w. post campus library as well as remotely. the layout of the information was designed to make the web site uncluttered and attractive. different color schemes were tried and informally polled among users. a version with white text on black background prompted strong likes or dislikes when shown to users. although this combination is easy to read, it was rejected because of the strong negative reactions from several users. photographs of the lis library and students were included. the pages were designed with a menu on the left side; fly-out menus were used to access submenus. where main library pages already existed for informa- tion to be included in the lis web site, such as lis hours and staff, links to those pages were made instead of re-cre- ating the information in the lis web site. an attempt was made to render the site accessible to users with disabilities, and pages were made compliant with the world wide web consortium (w c) by using their html validator and their cascading style sheet validator. ฀ literature review usability is a term with many definitions, varying by field. the fields of industrial engineering, product research and development, computer systems, and library science all share the study of human-and-machine interaction, as well user-centered design of a web site for library and information science students: heuristic evaluation and usability testing laura manzari and jeremiah trinidad-christensen laura manzari (manzari@liu.edu) is an associate professor and library and information science librarian at the c. w. post campus of long island university, brookville, n.y. jeremiah trinidad-christensen (jt @columbia.edu) is a gis/map librarian at columbia university, new york, n.y. information technology and libraries | september as a commitment to users. dumas and reddish explain it simply: “usability means that the people who use the product can do so quickly and easily to accomplish their own tasks.” user-centered design incorporates usability principles into product design and places the focus on the user during project development. gould and lewis cite three principles of user-centered design: an early focus on users and tasks, empirical measurement of product usage, and iterative design to include user input into product design and modification. jakob nielsen, an often-cited usability engineering specialist, emphasizes that for increased functionality, engineering usability principles should apply to web design, which should be treated as a software development project. he advocates incorporating user evaluation into the design process first through a heuristic evaluation, fol- lowed by usability testing with a redesign of the product after each phase of evaluation. usability principles have been applied to library web-site design; however, library web-site usability studies often do not include the addi- tional heuristic evaluation recommended by nielsen. in addition to usability, consideration should also be given during the design process to making the web site accessible to people with disabilities. federal agencies are now required by the rehabilitation act to make their web sites accessible to the disabled. section part . of the act enumerates sixteen rules for internet applications to help ensure web-site access for people with various dis- abilities. similarly, the web accessibility initiative hosted by the w c works to ensure that accessibility practices are considered in web-site design. they developed the web content accessibility guidelines for making web sites accessible to people with disabilities. although articles have been written about usability testing of academic library web sites, very little has been written about usability testing of special-collection web sites for distinct user populations within larger academic settings. ฀ heuristic evaluation methodology heuristic evaluation is a usability engineering method in which a small set of expert evaluators examine a user interface for design problems by judging its compliance with a set of recognized usability principles or heuristics. nielsen developed a set of ten widely adopted usability heuristics (see sidebar). after studying the use of individual evaluators as well as groups of varying sizes, nielsen and molich recommend using three to five evaluators for a heuristic evaluation. the use of multiple experts will catch more flaws than a single expert, but using more than five experts does not produce greater results. in comparisons of heuristic evaluation and usability testing, the heuristic evaluation uncovered more of the minor problems while usability test- ing uncovered more major, global problems. since each method tends to uncover different usability problems, it is recommended that both methods be used complementa- rily, particularly with an iterative design change between the heuristic evaluation and the usability testing. for the heuristic evaluation, four people were approached from the palmer lis school faculty and ph.d. program with expertise in web-site design and human- computer interaction. three agreed to participate. they were asked to familiarize themselves with the web site and evaluate it according to nielsen’s ten heuristics, which were provided to them. ฀ heuristic evaluation results the evaluators were all in agreement that the language was appropriate for lis students. one evaluator said if new students were not familiar with some of the terms they soon would be. another thought jake, the tool to access full text, might not be clear to students at first, but the lis web-site explanation was fine the way it was. they were also in agreement that the web site was well designed. comments included: “the purpose and descrip- tion of each page is short and to the point, and there is a good, clean, viewable page for the users”; “the site was well designed and not over designed”; “very clear and user friendly”; “excellent example of limiting unnecessary irrelevant information.” the only page to receive a “poor layout” comment was the lengthy subject list of journals, though no suggestions for improvement were made. concern was expressed about links to other web sites on campus. one evaluator thought new students might be confused about the relationship between long island university, c. w. post, and the palmer school. two evalua- tors thought links to the main library’s web site could cause confusion because of the different design and layout. a preference for the design of the lis library web site over the main library and palmer school web sites was expressed. to eliminate some confusion, the menu options for other cam- pus web sites were dropped down to a separate menu right below the menu of lis web pages. for additional clarity, some of the main library pages were re-created in the style of the lis pages instead of linking to the original page. the evaluators made several concrete suggestions for menu changes, which were included in the redesign. it was suggested that several menu options were unclear and needed clarification, so additional text was added for clarity at the expense of brevity. long island university’s online catalog is named liucat and was listed that way on the menu. new students might not be familiar with this name, so the menu label was changed to liucat (library catalog). user-centered design of a web site | manzari and trinidad-christensen for the link to jake, a description, find periodicals in online databases, was added for clarification. it was also suggested that the link to the main library web page for all databases could cause confusion since the layout and design of that page is different. the wording was changed to all databases (located in the c. w. post library web site). menu options were originally arranged in order of anticipated use (see figure ). thus, the order of menu options from the lis home page was databases, journals, library catalog, other web sites, palmer school, and main library. evaluators suggested that putting the option for lis home page first would give users an easy “emergency exit” to return to the home page if they were lost. the original menu options also varied from page to page. for example, menu options on the database page referred only to pages that users might need while doing database searches. at the suggestion of eva- luators, the menu options were changed to be con- sistent on every page (see figure ). a redesign based on these results was com- pleted and posted to the internet for public use (see figure ). ฀ usability testing methodology usability testing is an em- pirical method for improv- ing design. test subjects are gathered from the popu- lation who will use the product and are asked to perform real tasks using the prototype while their performance and reactions to the product are observed and recorded by an inter- viewer. this observation and recording of behav- ior distinguishes usability testing from focus groups. observation allows the tes- ter to see when and where users become frustrated or confused. the goal is to jakob nielsen’s usability heuristics visibility of system status—the system should always keep users informed about what is going on, through appropriate feedback within reasonable time. match between system and the real world— the system should speak the user’s language, with words, phrases, and concepts familiar to the user rather than system-oriented terms. follow real-world conventions, making information appear in a natural and logical order. user control and freedom—users often choose system functions by mistake and will need a clearly marked “emergency exit” to leave the unwanted state without having to go through an extended dialogue. support undo and redo. consistency and standards—users should not have to wonder whether different words, situations, or actions mean the same thing. follow platform conventions. error prevention—even better than good error messages is a careful design that prevents problems from occurring in the first place. recognition rather than recall—make objects, actions, and options visible. the user should not have to remember information from one part of the dialogue to another. instructions for use of the system should be visible or easily retrievable whenever appropriate. flexibility and efficiency of use—accelerators, unseen by the novice user, may often speed up the interaction for the expert user such that the system can cater to both inexperienced and experienced users. allow users to tailor frequent actions. aesthetic and minimalist design—dialogues should not contain information that is irrelevant or rarely needed. every extra unit of information in a dialogue competes with the relevant units of information and diminishes their relative visibility. help users recognize, diagnose, and recover from errors—error messages should be expressed in plain language (no codes), precisely indicate the problems, and constructively suggest a solution. help and documentation—even though it is better if the system can be used without documentation, it may be necessary to provide help and documentation. any such information would be easy to search, focused on the user’s task, list concrete steps to be carried out, and not be too large. figure . original menu figure . revised menu information technology and libraries | september uncover usability problems with the product, not to test the participants themselves. the data gathered are then analyzed to recommend changes to fix usability problems. in addition to recording empirical data such as number of errors made or time taken to complete tasks, active intervention allows the interviewer to question participants about reasons for their actions as well as about their opinions regarding the product. in fact, subjects are asked to verbalize their thought processes as they complete the tasks using the interface. test subjects are usually interviewed individually and are all given the same pretest briefing from a script with a list of instructions followed by tasks representing actual use. test subjects are also asked questions about their likes and dislikes. in most situations, payment or other incentives are offered to help recruit subjects. four or five subjects will reveal percent of usability problems. messages were sent to students via the palmer school’s mailing lists requesting volunteers. a ten-dollar gift certifi- cate to a bookstore was offered as an inducement to recruit- ment. input was desired from both master’s degree and doctoral students. the first nine volunteers to respond—all master’s degree students—were accepted. this group included students from both the main and satellite campuses. no ph.d. stu- dents volunteered to participate at first, citing busy schedules, but eventually a doctoral student was recruited. testing was conducted in computer labs at the library, at the palmer school, and at the manhattan satellite campus. demographic information was gathered regarding users’ gender, age range, university status, familiarity with computers, with the internet, and with the lis library, as well as the type of internet connection and browser usually used. the subjects were given eight tasks to complete using the web site. the tasks reflected both the type of assignment a student might receive in class and the type of information they might seek on the lis web site on their own. the questions were designed to test usability of different parts of the web site. ฀ ฀usability testing results the first task tested the print journals page and asked if the lis library subscribes to a specific journal and whether it is refereed. (the web site uses an asterisk next to a journal title to indicate that it is refereed.) all subjects were able to easily find that the lis library does hold the journal title. although it was not initially obvious that the asterisk was a notation indicating that the journal was refereed, most of the subjects eventually found the explanatory note. many of the subjects did not know what a refereed journal was, and some asked if a definition could be provided on the site. for the second task, subjects needed to use jake to find the full text of an article. none of the students were familiar with jake but were able to use the lis web site to gain an understanding of its purpose and to access it. the third task asked subjects to find a library asso- ciation that required using the other web sites page. all subjects demonstrated an understanding of how to use this page and found the information. the fourth task tested the full-text databases page. only one subject actually used this page to complete the task. the rest used the all databases link to the main library’s database list. that link appears above the link to full-text databases and most subjects chose that link without looking at the next menu option. several sub- figure . final home page user-centered design of a web site | manzari and trinidad-christensen jects became confused when they were taken to the main library’s page, just as the evaluators had predicted. even though wording was added warning users that they were leaving the lis web site, most subjects did not read it and wondered why the page layout changed and was not as clear. they also had trouble navigating back to the lis web site from the main library web site. the fifth task tested the journals by subject page. this task took longer for most of the subjects to answer, but all were able to use the page successfully to find a journal on a given subject. the sixth task required using the lis home page, and everyone easily used it to find the operating hours. the seventh task required subjects to find an online journal title that could be accessed from the electronic journals page. all subjects navigated this page easily. the final task asked subjects to find a book review. most subjects did not look at the page for library and information sciences databases to access the books in print database, saying they did not think it would be included there. instead, they used the link to the main library’s database page. one subject was not able to complete this task. problems primarily occurred during testing when sub- jects left the lis page to use a non-library science database located on the main web site. subjects had problems get- ting back to the lis site from the main library site. while performing tasks, some subjects would scroll up and down long lists instead of using the toolbars provided to bring the user to an exact location on the page. some preferred using the back button instead of using the lis web-site menu to navigate. these seemed to be individual styles of using the web and not any usability problem with the site. several people consistently used the menu to return to the lis home page before starting each new task, even though they could have navigated directly to the page they needed, making a return to the home page unnecessary. this validated the recommendation from the heuristic study that the link to the home page always be the first menu option to give users a comfortable safety valve when they get lost. the final questions asked subjects for their opinions on what they did and did not like about the web site, as well as any suggestions for improving the site. all subjects responded that they liked the layout of the pages, calling them uncluttered, clean, attractive, and logical. there were very few suggestions for improving the site. one person asked that contact information be included on the menu options in addition to its location right below the menu on the lis home page. another participant suggested adding class syllabi to the web site each semester, listing required texts along with a link to an online bookstore. some of the novice users asked for explanations of unfamiliar terms such as “refereed journals.” a participant suggested including a search engine instead of using links to navi- gate the site. this was considered during the initial site design but was not included since the site did not have a large number of pages. however, a search engine may be worth including. the one doctoral student had previously only used the main library’s web page to access databases. originally, he said he did not see the advantage of a site devoted to information science sources for doctoral candidates, since that program is more multidisciplinary. however, after completing the usability study, the student concluded that the lis web site was useful. he suggested that it should be publicized more to doctoral candidates and that it be more prominently highlighted on the main library web site. though the questions asked were about the lis web site, several subjects complained about the layout of the main library web site and suggested that it have better linking to the lis web site to enable it to be accessed more easily. ฀ conclusions iterative testing and user-centered design resulted in a product that testing revealed to be easy to learn and effi- cient to use, and about which subjects expressed satisfac- tion. based on findings that some students had not even been aware of the existence of the lis web site, greater emphasis is now given to the web site and its features during new student orientations. the biggest problem users had was navigating from the web pages of the main library back to the lis site. it was suggested that the lis site be highlighted more prominently on the main library web site. some users were confused by the different layouts between the sites, but no one expressed a preference for the design used by the main library web site. despite this confusion, subjects overwhelmingly expressed positive feedback about having a specialized library site serving their specific needs. issues regarding web-site design can be problematic for smaller specialized libraries within larger institutions. in this case, some of the problems navigating between the sites could be resolved by changes to the main library site. the design of the lis web site was preferred over the main campus web site by both the heuristic evaluators and the students in the usability test. however, designers of a main library web site might not be receptive to suggestions from a specialized or branch library. although consistency in design would eliminate confusion, requiring the special- collection’s web site to follow a design set by the main institution could be a loss for users. in this instance, the main site was designed without user input, whereas the specialized library serving a smaller population was able to be more dynamic and responsive to its users. finding an appropriate balance for a site used by students new to the field as well as advanced students is information technology and libraries | september a challenge. although the students in the study were all experienced computer and web users, their familiarity with basic library concepts varied greatly. a few novice users expressed some confusion as to the difference between journals and index databases. there actually was a description of each of these sources on the site but it was not read. (the subjects barely read any of the site’s text, so it can be difficult to make some points clearer when users want to navigate quickly without reading instructions. several subjects who did not bother to read text on the site still suggested having more notes to explain unfamiliar terms. however, if the site becomes too overloaded with explanations of library concepts, it could become annoying for more advanced users.) a separate page with a glos- sary is a possibility—based on the study, however, it will probably not be read. another possibility is a handout for students that could have more text for new users without cluttering the web site. having such a handout would also serve to publicize the site. there was some concern prior to the study that offer- ing more advanced features, such as providing access to jake or indicating which journals are refereed, might be off-putting for new students; therefore, test questions were designed to gauge reactions to these features. most students in the study did express some intimidation at not being familiar with these concepts. however, all the subjects eventually figured out how to use jake and, once they tried it, thought it was a good idea to include it. even new students who had the most difficulty were still able to navigate and learn from the site to be able to use it efficiently. an online survey was added to the final design to allow continuous user input. the site consistently receives posi- tive feedback through these surveys. it was planned that responses could be used to continually assess the site and ensure that it is kept responsive and up-to-date; however specific suggestions have not yet been forthcoming. how valuable was usability testing to the web-site design? several good suggestions were made and imple- mented, and the process confirmed that the site was well designed. it provided some insight into how subjects used the web site that had not been anticipated by the design- ers. since usability studies are fairly easy and inexpensive to conduct, it is probably a step worth taking during the web-site design process even if it results in only minor changes to the design. references and notes . w c, “the w c markup validation service,” validator .w .org (accessed nov. , ); w c, “the w c css validation service,” jigsaw.w .org/css-validator (accessed nov. , ). . see carol m. barnum, usability testing and research (new york: longman international, ); alison j. head, “web redemption and the promise of usability,” online , no. ( ): – ; international standards organization, ergonomic requirements for office work with visual display terminals. part : guidance on usability—iso - (geneva: international organization for standardization, ); judy jeng, “what is usability in the context of the digital library and how can it be measured?” information technology and libraries , no. ( ): – ; jakob nielsen, usability engineering (boston: academic, ); ruth ann palmquist, “an overview of usability for the study of users’ web-based information retrieval behavior,” journal of education for library and information science , no. ( ): – . . joseph s. dumas and janice c. redish, a practical guide to usability testing (portland: intellect bks., ), . . john d. gould and clayton h. lewis, “designing for usability: key principles and what designers think,” commu- nications of the acm no. ( ): – . . jakob nielsen, “heuristic evaluation,” in jakob nielsen and robert l. mack, eds., usability inspection methods (new york: wiley, ), – . . see denise t. covey, usage and usability assessment: library practices and concerns (washington, d.c.: digital library federation, ); nicole campbell, usability assessment of library-related web sites (chicago: ala, ); kristen l. garlock and sherry piontek, designing web interfaces to library services and resources (chicago: ala, ); anna noakes schulze, “user-centered design for information professionals,” journal of education for library and information science , no. ( ): – ; susan m. thompson, “remote observation strategies for usability testing,” information technology and libraries , no. ( ): – . . government services administration, “section : sec- tion standards,” www.section .gov/index.cfm?fuseacti on=content&id= #web (accessed nov. , ). . w c, “web content accessibility guidelines . ,” www .w .org/tr/wcag (accessed nov. , ). . see susan augustine and courtney greene, “discover- ing how students search a library web site: a usability case study,” college and research libraries , no. ( ): – ; brenda battleson, austin booth, and jane weintrop, “usability testing of an academic library web site: a case study,” journal of academic librarianship , no. ( ): – ; janice krueger, ron l. ray, and lorrie knight, “applying web usability tech- niques to assess student awareness of library web resources,” journal of academic librarianship , no. ( ): – ; thura mack et al., “designing for experts: how scholars approach an academic library web site,” information technology and librar- ies , no. ( ): – ; mark shelstad, “content matters: analysis of a web site redesign,” oclc systems & services , no. ( ): – ; robert l. tolliver et al., “web site redesign and testing with a usability consultant: lessons learned,” oclc systems & services , no. ( ): – ; dominique turnbow et al., “usability testing for web redesign: a ucla case study,” oclc systems & services , no. ( ): – ; leanne m. vandecreek, “usability analysis of northern illinois user-centered design of a web site | manzari and trinidad-christensen university libraries’ web site: a case study,” oclc systems & services , no. ( ): – . . jakob nielsen and rolf molich, “heuristic evaluation of user interfaces,” in proceedings of the acm chi ’ (new york: association for computing machinery, ), – . . robin jeffries et al., “user interface evaluation in the real world: a comparison of a few techniques,” in proceedings of the acm chi ’ (new york: association for computing machin- ery, ), – ; jakob nielsen, “finding usability problems through heuristic evaluation,” in proceedings of the acm chi ’ (new york: association for computing machinery, ), – . . jakob nielsen, “heuristic evaluation,” – . . jeffrey rubin, handbook of usability testing: how to plan, design, and conduct effective tests (new york: wiley, ); jakob nielsen, “why you only need to test with five users, alert- box mar. , ,” www.useit.com/alertbox/ .html (accessed nov. , ). information technology and libraries | december social engineering is the use of non- technical means to gain unauthorized access to information or computer systems. while this method is rec- ognized as a major security threat in the computer industry, little has been done to address it in the library field. this is of particular concern because libraries increasingly have access to databases of both proprietary and personal information. this tutorial is designed to increase the awareness of library staff in regard to the issue of social engineering. one morning the phone rings at the circulation desk; the assistant, joyce, answers. “seashore branch public library, how may we help you?” she asks, smiling. “my wife and i recently moved and i wanted to confirm that you had our current address,” a pleas- ant male voice responds. “could you give me your name please?” “the card is in my wife’s name, jennifer greene. we’ve been so busy with the move that she hasn’t had a chance to catch up with everything.” “okay, i have her information here. main street, apartment b. is that correct?” “thank you so much, that’s it. do you have our new number or is it still - - in your records?” “let me see . . . no, i think we have your new number.” “could you read it back to me?” “sure . . . - - , is that right?” “ - - . . . that’s right. thank you very much, you’ve been very helpful.’ “no problem, that’s what we’re here for.” what just happened? what happened to joyce may have been exactly what it appeared to be—a conscientious spouse trying to make sure information was updated after a move. but what else could it have been—research for an identity theft, or a stalker trying to get personal information? we have no way of knowing. all reasons except for the first, innocent, reason are covered by the term social engineering. in the language of computer hackers, social engineering is a non- technical hack. it is the use of trickery, persuasion, impersonation, emotional manipulation, and abuse of trust to gain information or computer-system access through the human interface. regardless of an institution’s commit- ment to computer security through technology, it is vulnerable to social engineering. recently, the institute of management and administration (ioma) reported social engineering as the number-one security threat for . according to ioma, this method of security violation is on the rise due to continued improvements in techni- cal protections against hackers. why and how does social engineering work? the first thing to keep in mind about social engineering is that it does work. kevin mitnick, possibly the best known hacker of recent decades, carried out most of his questionable activities through the medium of social engineering. he did not need to use his technical expertise because it was easier to just ask for the infor- mation he wanted. he discovered that people, when questioned appropri- ately, would give him the information he wanted. social engineering succeeds because most people work under the assumption that others are essentially honest. as a pure matter of probabil- ity, this is true; the vast majority of communications that we receive dur- ing the day are completely innocent in character. this fact allows the social engineer to be effective. by making seemingly innocuous requests for information, or making requests in a way that seems reasonable at the time, the social engineer can gather the information that he or she is look- ing for. methods of social engineering the arsenal of the social engineer is large and very well established. this is mainly because social engineering amounts to a variation on confidence trickery, an art that goes back as far as human history can recall. one might argue that homer’s iliad contains the first record of a social engineer- ing attack in the form of the trojan horse. direct requests many social-engineering methods are complex and require significant plan- ning. however, there is a simple and effective method that is often just as effective. the social engineer contacts his or her target and simply asks for the information. preying on trust and emotion social engineering is a method of gain- ing information through the persua- sion of human sources, based on the abuse of trust and the manipulation of emotion. in his book, the art of deception, mitnick makes the argu- ment that once a social engineer has established the trust of a contact, then all security is effectively voided and helping the hacker? library information, security, and social engineering samuel t. c. thompson samuel t. c. thompson (sthompson@ collier-lib.org), is a public service librar- ian at the collier county public library, naples, florida. helping the hacker? | thompson the social engineer can gather what- ever information is required. the most common method of tar- geting computer end-users is through the manipulation of gratitude. in these cases, a social engineer, usually impersonating a technician, contacts a user and states that there is something wrong on the victim’s end, and that the social engineer needs a few pieces of information to “help” the user. appreciative of the assistance, the vic- tim provides the necessary informa- tion to the helpful caller or carries out the requested actions. predictably, no problem ever existed and the victim has now provided the social engineer either access to a computer system or with the information needed to gain that access. a counterpoint to the manipula- tion of gratitude is the manipulation of sympathy. this method is most often used on information providers such as help-desk personnel, techni- cians, and library staff members. in this scenario, a social engineer con- tacts a victim and claims to have either lost information, is out of contact with a normal source, or is simply ignorant of something that he or she should know. as anyone can empathize with this plea, the victim is often all too willing to provide the information sought by the social engineer. using these methods—taking advantage of the gratitude, sympathy, and empathy of their victims—social engineers are able to achieve their aims. impersonation because forming trust relationships with their victims is critical to a social- engineering attack, it is not surprising that social engineers often pretend to be someone or something that they are not. two of the major tools of imper- sonation are ( ) speaking the language of the victim institution and ( ) knowl- edge of personnel and policy. to allay suspicion, a social engi- neer needs to know and be able to use an institution’s terminology. being unable to do so would cause the victim to suspect, rather than trust, the social engineer. with a working knowledge of an organization’s par- ticular vocabulary, a social engineer can phrase his or her request in terms that will not rouse alarm with the intended victim. the other major goal of a social engineer in preparing a successful impersonation is to develop a famil- iarity with the “lay of the land,” i.e., the specifics of and personnel within an organization. for instance, a social engineer needs to discover who has what authority within an organization so as to understand for whom he or she needs to claim to speak. research to establish trust in their victims, social engineers use research as a tool. this comes in two forms, background research and cumulative research. background research is the pro- cess by which a social engineer uses publicly available resources to learn what to ask for, how to ask for it, and whom to ask it of. while the intent and goal of this research differs from the techniques used by students, librarians, and other members of the population, the actual process is the same. cumulative research is the process by which a social engineer gathers the information that he or she needs to make more critical requests of their victims. the facts that a social engineer seeks through cumulative research may seem without value to the casual observer, but put together properly, they are anything but that. questions can include names of staff, internal phone numbers, procedures, or seemingly minor technical details about the library’s network (e.g., what operating system are you running?). late in the afternoon the phone at the reference desk rings. marcy, the librarian on duty answers, “reference desk.” “hi there, this is dave simpson calling from information services at the main branch. sorry about the echo, i’m working in the cabling closet at the moment, so i’m calling you on my cell phone.” “no problem, i can hear you fine. what can i do for you?” “thanks. a lot of the branches have been having network problems over the last few days. has everything been okay at the seashore branch reference desk?” “i think so.” “okay, that’s good. i’m running a test right now on the network and needed to find a terminal that was behaving itself. could you log off and let me know if any messages come up?” “no problem.” marcy logs off of the reference computer; nothing strange happens. “just the usual mes- sages.” “good. now start logging back on. what user are you going in as? i mean which login name are you using?” “searef. okay, i’m logged on now.” “no strange messages?” “nothing.” “that’s great. look, our problem might be kids hacking into the system so i need you to change the password. do you know how to do that?” “i think so.” “well, let me walk you through it.” dave spends a couple of minutes walking marcy through changing the system password. the password is now changed to ear f, a moderately secure password. “thanks, marcy. you’ve been a great help. we have your new password logged into the system. could you pass on the new password to the other reference per- sonnel?” “sure.” “wonderful. just remember not to give the password out to anyone who doesn’t need it, and don’t write it down where anyone who shouldn’t have it can get at it. have a great day.” “you too.” information technology and libraries | december why are libraries vulnerable? libraries are vulnerable to social-engi- neering attacks for two major reasons: ( ) ignorance and ( ) institutional psychology. the first of these diffi- culties is the easiest to address. the ignorance of library professionals in this matter is easily explained—there is very little literature to date about the issue of social engineering directed at library personnel. what exists is usually mixed in larger articles on general security issues and receives little focus. this lack of concern about social engineering can also be seen in com- puter professional literature, where it is dwarfed by the volume of articles concerning technical security issues. this is a curious gap, considering the high rate of occurrence of this kind of attack. is it because many techni- cal professionals are less comfortable with a social issue—that can only be solved through people—than with a technical security issue that can be solved through the development or implementation of proper software? unfortunately, not knowing about a method of security violation leaves one vulnerable to that method. it is incumbent on librarians, computer administrations, and security profes- sionals to be aware of these issues. the second factor is harder to address but equally important. unlike almost any other profession, librarians are expected to fulfill their patrons’ informational needs without ques- tion or bias. this laudable goal makes librarians vulnerable to social-engi- neering attacks because the inquiries made by a social engineer about the information resources available at a library may be used for nefarious purposes. a reference interview over these issues may be very successful from the point of view of both parties involved, as the librarian fills the open- ended inquiries of the social engineer, and the social engineer receives much, if not all, of the information that he or she needs to violate the library’s internal information systems. why libraries can be targets at this point, it is relevant to ask why security violators would even bother with library computer networks. what do libraries have that is worth possibly committing a crime to get? personal information is probably the most tempting target in a library computer system. libraries possess databases of names, addresses, and other personal data about library card- holders. this information is valuable, and not all of it is easily available from public sources. as may be seen in the section of this article on techniques, such information could be used as an end unto itself or as a stepping stone to security violations in other systems. subscriptions to proprietary data- bases are quite expensive, as any acquisitions librarian will explain. given the high prices and limited licensing, a hacker may want to gain access to these information resources. this could be a casual hacker who wants to have access to a library-only resource from his or her home com- puter, or this may be a criminal who wishes to steal intellectual properties from a database provider. libraries often have broadband access designed for a large network (e.g., t ). as these lines are very expensive, few individuals can afford them. at the same time, it has been observed that these broadband lines have immense capabilities for down- loading information from other net- works. there are many reasons why a hacker would seek to illicitly use such a resource. for instance, a casual hacker may want to download a large number of bootlegged movie files, or a criminal may wish to download a corporate database. with access to a library’s high bandwith internet line, these actions can be carried out quickly and with a minimized chance of detection. libraries possess large numbers of computers due to their increas- ing automation. these computer resources can, if compromised, be used as anonymous remote comput- ers by hackers. called “zombies,” compromised computers could be used to deliver illegal spam, distrib- uted denial of service (ddos) attacks, or as servers to distribute illegal materials. if library computers are used in this way, there is a potential for a library to face legal responsibil- ity for the actions of its computers or for the questionable materials found on them. prevention the tools needed to prevent social engineering from succeeding are awareness, policy, and training. these tools feed into one another—we become aware of the possibility of social-engineering attacks, develop policy to communicate these concerns to others, and then train others in these policies to protect them and their libraries from social engineering. libraries should have a simple set of policies to help prevent social engi- neering from affecting them. this pol- icy need not be long; ideally, it should be a small page of bullet points that are easy to remember or to post near telephones. what is important is that it is easy to remember and implement when a call or e-mail comes in. basic guidelines for protection against social engineering ■ be suspicious of unsolicited communications asking about employees, technical informa- tion, or other internal details. ■ do not provide passwords or login names over the phone or helping the hacker? | thompson via e-mail no matter who claims to be asking. ■ do not provide patron informa- tion to anyone but the patron in person and only upon presenta- tion of the patron’s library card or other proper identification. ■ if you are not sure if a request is legitimate, contact the appropri- ate authorities. ■ trust your instincts. if you feel suspicious about a question or communication, there is prob- ably a good reason. ■ document and report suspicious communications. in closing social engineering is an immensely effective method of breaching com- puter and network security. it is, how- ever, entirely dependent on the ability of the social engineer to persuade staff members into providing information or access that they should not provide. with care and good information poli- cies, we can prevent social engineer- ing from working. after all, do we really want to be helping the hacker? the circulation desk phone rings. joyce answers, “seashore branch public library, how may we help you?” “hi there, i’m worried that i haven’t turned in all the books i have out, and i really don’t want to get stuck with a fine. could you tell me what i have out?” “no problem. what is you name?” “sean grey.” joyce brings up sean grey’s circu- lation records, and then remembers about the library’s information policy and decides to ask another question, “could you give me your library card number?” “i don’t have that with me. i really don’t want to get stuck with those fines.” “i’m sorry. mr. grey, to preserve patron privacy we can only give out circulation information if you give us your card number or if you are here in person with your card or id.” “but i just want to avoid a fine. can’t you help?” “don’t worry; if you are late by accident on occasion, we are willing to forgive a fine.” “so you can’t give me my records?” “i’m sorry but we have to protect patron privacy. i’m sure you under- stand.” “i guess so. goodbye.” “have a good day.” ■ references . institute of management & admin- istration, “six security threats that will make headlines in ’ ,” ioma’s security director’s report , no. ( ): – . . k. manske, “an introduction to social engineering,” security management practices (nov./dec. ): – . . m. mcdowell, “cyber-security tip st - ,” ( ), http://www.us.cert. gov/cas/tips/st - .html (accessed june , ). . k. mitnick and w. simon, the art of deception (indianapolis: wiley, ). alcts cover lama cover lita , , cover index to advertisers