Bellinger Named Director of the Office of Digital Assets and Infrastructure at Yale

Meg Bellinger, Associate University Librarian for Integrated Access and Technical Services at the Yale University Library, has been named Director of the Office of Digital Assets and Infrastructure at Yale, a new position in the Provost's office that is responsible for university-wide digitization.

Read more about it at "Bellinger to Direct Digitizing Office."

Google Newspaper Digitization Project Announced

Google has announced a newspaper digitization project that will "make more old newspapers accessible and searchable online by partnering with newspaper publishers to digitize millions of pages of news archives."

Read more about it at "Bringing History Online, One Newspaper at a Time."

Technical Report: Doctoral Theses Digitisation

Ingrid Mason, Digital Research Repository Coordinator at the New Zealand Electronic Text Centre of the Victoria University of Wellington’s University Library, has deposited a report (Technical Report: Doctoral Theses Digitisation) about that library's doctoral theses digitization project in its institutional repository.

Here's an excerpt:

Doctoral theses (~1200) in the University Library’s collection have been digitised and uploaded into the Library’s two research repositories: RestrictedArchive@Victoria and ResearchArchive@Victoria. With a view to sharing learning and useful information key considerations for other tertiary institutions undertaking a similar project are:

  • digital file sizes and server storage space
  • purpose of and standards of digitisation for access
  • data matching from library system and alumni database
  • database listing and tracking of theses and allied tasks
  • inventory listing and batching of theses into boxes
  • costs for digitisation, transportation and short term assistance

British Library Releases Its "Digitisation Strategy 2008-2011"

The British Library has released its "Digitisation Strategy 2008-2011."

Here's an excerpt:

Over the next 3 years we will build on our existing digitisation programme. Current projects include the digitisation of:

  • 20 million pages of 19th century literature [approximately 80,000 books];
  • 1 million pages of historic newspapers in addition to the 3m already digitised;
  • 4,000 hours of Archival Sound Recordings in addition to the 4,000 hours already digitised;
  • 100,000 pages of Greek manuscripts.

Our top priority digitisation programme in support of the Library's corporate strategy 2008-2011 is the digitisation of newspapers.

OCLC Announces WorldCat Copyright Evidence Registry Beta

OCLC has announced the WorldCat Copyright Evidence Registry beta, a union catalog of copyright information.

Here's an excerpt from the press release:

The WorldCat Copyright Evidence Registry is a community working together to build a union catalog of copyright evidence based on WorldCat, which contains more than 100 million bibliographic records describing items held in thousands of libraries worldwide. In addition to the WorldCat metadata, the Copyright Evidence Registry uses other data contributed by libraries and other organizations.

Digitization projects continue for books in the public domain, but books whose copyright status is unknown are destined to remain in print and on shelves until their status can be determined. The process to determine copyright status can be lengthy and labor intensive. The goal of the Copyright Evidence Registry is to encourage a cooperative environment to discover, create and share copyright evidence through a collaboratively created and maintained database, using the WorldCat cooperative model to eliminate duplicate efforts. . . .

The Copyright Evidence Registry six-month pilot was launched July 1 to test the concept and functionality. Users can search the Copyright Evidence Registry to find information about a book, learn what others have said about its copyright status, and share what they know. . . .

During a later stage of the pilot, OCLC will add a feature enabling pilot libraries to create and run automated copyright rules conforming to standards they define for determining copyright status. The rules will help libraries analyze the information available in the Copyright Evidence Registry and form their own conclusions about copyright status.

Five TexTreasures Digitization Grants Awarded

The Texas State Library and Archives Commission has awarded digitization grants to five TexShare member libraries.

Here's an excerpt from the press release:

TSLAC received 28 TexTreasures grant proposals. The exciting projects that have been funded are:

  1. "Houston Oral History Project" ($17,474)—The Houston Public Library is partnering with Mayor Bill White to preserve and make the video-recordings of significant Houstonians available on the web.
  2. "Early Texas Newspapers: 1829-1861" ($24,637)—The University of North Texas Libraries and the Center for American History at the University of Texas at Austin will partner to microfilm, digitize, and provide free public access to the earliest Texas newspapers held by the Center for American History.
  3. "The Witliff Collections" ($20,000)—The project creates an online exhibit accessing the primary source materials of researcher Dick J. Reavis held by the Southwestern Writers Collection at the Wittliff Collections at Texas State University about the siege of the Branch Davidians at Mount Carmel outside of Waco in 1993.
  4. "Austin History Center Glass Plate Negatives" ($12,889)—The Austin History Center, a division of the Austin Public Library, will digitize the complete Hubert Jones collection of 471 glass plate negatives containing subjects local to Austin and Texas.
  5. "Tejano Voices Project" ($20,000)—The University of Texas at Arlington Library will digitize and describe 60 of the 174 oral history interviews with notable Tejanos and Tejanas from across Texas conducted in 1992-2003 by Dr. Jose Angel Gutierrez, associate professor of political science at UT Arlington.

The Impact of Digitizing Special Collections on Teaching and Scholarship: Reflections on a Symposium about Digitization and the Humanities

OCLC Programs & Research has released The Impact of Digitizing Special Collections on Teaching and Scholarship: Reflections on a Symposium about Digitization and the Humanities.

Here's an excerpt:

University faculty and scholars demonstrated their uses of rare books and archives—in both digital and physical forms—to an audience of RLG Programs partners at a symposium in Philadelphia on June 4, 2008. Tony Grafton's recent article in The New Yorker provoked the theme of the symposium: we'll be travelling both the wide smooth road through the screen and the narrow difficult road of books and archives for a long time to come.

The audience of librarians, archivists, museum professionals and senior managers discussed administrative issues and opportunities for the use of digitized special collections. The academic speakers, however, spoke to us directly about their expectations of special collections and proposals for collaboration with scholars. These scholars emphasized the critical roles rare books, archives and other materials play in both teaching and research, and called for specific directions for libraries and archives to take in the near future. The primary users of primary resources presented clear imperatives for collections and custodians: work with faculty to understand current research methods and materials; go outside the library or archive to build collections and work with faculty; and continue to build digital and material collections for both teaching and research.

National Center for Research in Advanced Information and Digital Technologies to Be Established

A National Center for Research in Advanced Information and Digital Technologies will be established as part of the Higher Education Opportunity Act (see Sec. 802).

Here's an excerpt from the Digital Promise press release on its home page:

The new program is entitled the "National Center for Research in Advanced Information and Digital Technologies." It is a Congressionally originated 501(c)(3) nonprofit corporation located within the Department of Education. It will have a nine-member independent Board of Directors appointed by the Secretary of Education from nominations by members of Congress. Grants and contracts will be awarded on merit, and policies will be developed following the tested procedures of NSF and NIH. Given its status as a non-profit, independent corporation, the Center will be able to receive grants, contracts, and philanthropic contributions, as well as federal appropriations. . . .

Our next challenge is to secure FY09 appropriations for the Center. Because of the delay in passing the Higher Education Act, it was not possible for appropriations of the, until now, unauthorized National Center to be included in the Labor, HHS or Education funding bills that were passed in Committee in June. It is widely expected that final appropriations for FY09 will not be enacted until early next year. We are working hard to have funding for the National Center included in final appropriations legislation. We are requesting $50 million for FY09.

According to the About Digital Promise page one of the functions of the center will be to "commission pre-competitive research and fund the development of prototypes to . . . Digitize America’s collected memory stored in our nation’s universities, libraries, museums and public television archives to make these materials available anytime and anywhere."

A Look at the British Library's Digitization Program's Copyright Challenges

Tracey Caldwell's recent "Scan and Deliver" article examines the copyright challenges that the British Library faces in its digitization program (e.g., copyright issues have to be considered for works going as far back as the 1860s). It also mentions the impact of the shutdown of Microsoft's book digitization program on the British Library (digitization costs were shared 50-50 with Microsoft).

Research Study: How Is Web 2.0 Viewed by Academics?

The Birmingham Museums and Art Gallery's Pre-Raphaelite digitization project has released a study (Pre-Raphaelite Resource Project: Audience Research Report) about the perceptions of academics of the usefulness of Web 2.0 capabilities.

Here's an excerpt from the "Executive Summary":

Our research indicated that there is some readiness among the education community for Web 2.0 technologies but only in the context of academia as a status-conscious, competitive environment. Whilst there are clear benefits to be achieved from providing teachers and students with the opportunity to share ideas in the context of stimulus artefacts, many hold reservations about 'giving away' their intellectual property. Providing different levels of publishing privileges will help cater for the varying acceptance within the audience base for sharing their ideas publicly.

Social networking features are perceived by both HE students and lecturers as primarily for pleasure rather than for work so must be used sparingly in a resource of this nature. For younger students, however, the boundaries between work and life are increasingly blurred and the ability to contact experts and to personalise or control the space would be welcomed.

Care must be taken with positioning for the resource to be truly useful as a research tool; students and lecturers need to know that it has been created for them and has scholarly merit. Their main concern is to access reliable, relevant content and information, but the ability to form connections between these resources is one way of adding value to the collection.

Critique of the National Archives' The Founders Online Report

Peter Hirtle has posted a sharp critique of the National Archives' The Founders Online report on the LibraryLaw Blog that, among other points, questions whether the digitized works that result from the project will be free of copyright and access restrictions.

Here's an excerpt:

5. Perhaps the most problematic issues in the report surround its use of the term "open access." For some, open access means "digital, online, and free of charge." The report, while saying it wants to provide open access to the material, appears to recommend that all material be given to UVA's Rotunda system for delivery. Rotunda follows a subscription model—not open access—that is remarkably expensive considering that citizens have already paid for all of the editorial work on these volumes. How could this be open access? Apparently Rotunda might be willing to give up its subscription approach if a foundation were willing to pay for all of its costs. Unless such a commitment is in place, I find it disingenuous to describe a Rotunda delivery option as "open access." There is no discussion of other, free, delivery options, such as the willingness expressed by Deanna Marcum of the Library of Congress at the Senate Hearing to make all of the Founding Fathers papers accessible through LC (which already has a good site pointing to currently accessible papers).

6. Others argue that for true open access, information must be accessible outside of specific delivery systems (such as Rotunda) and made available in bulk. Open data and open interfaces allow for all sorts of interesting uses of material. For example, someone might want to mashup George Washington's papers to Google Maps in order to be able to easily visual geographically the spread of information. Others might want to mesh manuscript material with published secondary literature. Rather than anticipating the widespread dispersal and re-use of the Founding Fathers papers, however, and hence the need for harvestable data, open APIs, distributed access, etc., the report calls instead for "a single, unified, and sustainable Web site"—apparently the locked-down Rotunda system.

NEH/DFG Bilateral US/German Humanities Digitization Grants

The National Endowment for the Humanities (NEH) and the German Research Foundation (Deutsche Forschungsgemeinschaft) have issued a call for bilateral US/German humanities digitization grant proposals.

Here's an excerpt from the call:

These grants provide funding for up to three years of development in any of the following areas:

  • new digitization projects and pilot projects;
  • the addition of important materials to existing digitization projects; and
  • the development of related infrastructure to support international digitization work and the use of those digitized resources.

Collaboration between U.S. and German partners is a key requirement for this grant category.

Registry of U.S. Government Publication Digitization Projects Enhanced

The Registry of U.S. Government Publication Digitization Projects has been significantly enhanced.

Here's an excerpt from the announcement:

The enhanced Registry provides the ability to:

  • Browse digitization projects by category or alphabetically by title.
  • Search the entire Registry or filter searches by category or fields.
  • Quickly access new and recently updated listings.
  • Utilize RSS feeds to keep informed of new and updated projects.
  • View listings by contributor.
  • Contact fellow digitization participants.
  • Recommend listings to others.
  • Report broken links.
  • And much more!

Coverage of the Demise of Microsoft's Mass Digitization Project

Microsoft's decision to end its Live Search Books program, which provided important funding for the Open Content Alliance, has been widely covered by newspapers, blogs, and other information sources.

Here's a selection of articles and posts: "Books Scanning to be Publicly Funded," "'It Ain’t Over Till It's Over': Impact of the Microsoft Shutdown," "Microsoft Abandons Live Search Books/Academic Scan Plan," "Microsoft Burns Book Search—Lacks 'High Consumer Intent,'" "Microsoft Shuts Down Two of Its Google 'Wannabe’s': Live Search Books and Live Search Academic," "Microsoft Will Shut Down Book Search Program," "Microsoft's Book-Search Project Has a Surprise Ending," "Post-Microsoft, Libraries Mull Digitization," "Publishers Surprised by Microsoft Move," "Why Killing Live Book Search Is Good for the Future of Books," and "Without Microsoft, British Library Keeps on Digitizing."

University of Florida Has Digitized 1.7 Million Pages, over 100,000 in Last Month Alone

The University of Florida Digital Library Center has announced that it has digitized over 1.7 million pages, with about 100,000 pages being added in the last month alone. Their digitization statistics are available online. (Thanks to Open Access News.)

Read more about it "100,000 Pages a Month."

OCLC Announces Digital Archive Service

OCLC has announced the availability of a Digital Archive service.

Here's an excerpt from the press release:

The service provides a secure storage environment for libraries to easily manage and monitor master files and digital originals. The importance of preserving master files grows as a library's digital collections grow. Libraries need a workflow for capturing and managing master files that finds a balance between the acquisition of both digitized and born-digital content while not outpacing a library's capability to manage these large files. . . .

The Digital Archive service is a specially designed system in a controlled operating environment dedicated to the ongoing managed storage of digital content. OCLC has developed specific systems processes and procedures for the service tuned to the management of data for the long term.

From the time content arrives, the Digital Archive systems begin inspecting it to ensure continuity. OCLC systems perform quality checks and record the results in a "health record" for each file. Automated systems revisit these quality checks periodically so libraries receive up-to-date reports on the health of the collection. OCLC provides monthly updated information for all collections on the personal archive report portal.

For users of CONTENTdm, OCLC's digital collection management software for libraries and other cultural heritage institutions, the Digital Archive service is an optional capability integrated with various workflows for building collections. Master files are secured for ingest to the Digital Archive service using the CONTENTdm Acquisition Station, the Connexion digital import capability and the Web Harvesting service.

For users of other content management systems, the Digital Archive service provides a low-overhead mechanism for safely storing master files.

California Digital Library Puts Up Mass Digitization Projects Page

The California Digital Library has added a UC Libraries Mass Digitization Projects page to its Inside CDL Web site.

The Web site includes links to Frequently Asked Questions, contracts with digitization partners, and other information.

Of special interest in the FAQ are the questions "What rights to the digitized content does UC have in the projects; will access be limited in any way?" and "How will our patrons be able to access these texts, i.e. through MELVYL, or local catalogs, or a webpage, any search engine, or….?"

Ball State University Libraries Move Ahead with Ambitious Digital Initiative Program

The Ball State Libraries have nurtured an ambitious digital initatives program that has established an institutional repository, a CONTENTdm system for managing digital assets, a Digital Media Repository with over 102,000 digital objects, a Digitization Center and Mobile Digitization Unit, an e-Archives for university records, and a virtual press (among other initiatives). Future goals are equally ambitious.

Read more about it at "Goals for Ball State University Libraries' Digital Initiative."

RLG Program Releases Copyright Investigation Summary Report

OCLC's RLG Program has released the Copyright Investigation Summary Report.

Here's an excerpt from the announcement:

This report summarizes interviews conducted between August and September 2007 with staff RLG Partner institutions. Interviewees shared information about how and why institutions investigate and collect copyright evidence, both for mass digitization projects and for items in special collections.

Recipients of the JISC/NEH Transatlantic Digitization Collaboration Grants Announced

The new Office of Digital Humanities of the National Endowment for the Humanities announced the first five recipients of the of the JISC/NEH Transatlantic Digitization Collaboration Grants yesterday.

Here's an excerpt from the press release:

The announcement was made by NEH Chairman Bruce Cole during an event at the Folger Shakespeare Library. . . . A total of five projects received over $600,000 in funding. . . .

Inaugurated last year as part of the Endowment’s Digital Humanities Initiative, the JISC/NEH Transatlantic Digitization Collaboration Grant program is supported by both the NEH and the Higher Education Funding Council for England acting through the Joint Information Systems Committee (JISC). These grants provide combined funding of up to $240,000 for one year of development in the following areas: new digitization projects and pilot projects, the addition of important materials to existing digitization projects, or the development of infrastructure (either technical "middleware," tools, or knowledge-sharing) to support U.S.-England digitization work. Each project is sponsored by both an American and an English institution, whose activities will be funded by NEH and JISC respectively.

The formation of the Endowment’s Office of Digital Humanities (ODH) also was announced during the event. In 2006, the NEH launched the Digital Humanities Initiative, a program encouraging and supporting projects that utilize or study the impact of digital technology on research, education, preservation, and public programming in the humanities. With the creation of ODH, the initiative is being made permanent as an office within the NEH. ODH will continue the work of the initiative and will help to coordinate the Endowment’s efforts in the area of digital scholarship.

A complete list of the projects announced can be found below:

The Folger Shakespeare Library and the University of Oxford, with the Maryland Institute for Technology in the Humanities at the University of Maryland and the Shakespeare Institute at the University of Birmingham, plan to create the Shakespeare Quartos Archive, a freely-accessible, high-resolution digital collection of the seventy-five quarto editions of William Shakespeare's plays. The project will also develop an interactive interface and toolset for the detailed study of the quartos, with full-functionality applied to all thirty-two copies of one play, Hamlet, held at participating institutions, including the British Library, the University of Edinburgh Library, the Huntington Library, and the National Library of Scotland. ($119,598)

The Internet Archive in the United States and the Oxford Internet Institute and Hanzo in the United Kingdom plan to develop procedures and tools to improve the effectiveness of humanities research on the Web. This research and development effort promises to yield superior methods for indexing and analyzing the textual parts of larger digital collections, more focused browsing ("crawling") of the Web, and unified access to data resources, i.e., the ability to search for information across multiple digital databases. ($106,395)

The Institute for the Study of the Ancient World at New York University and the Centre for Computing in the Humanities at King's College, London, plan to launch Concordia, a set of tools and procedures to enable seamless textual searches and the dynamic mapping of a variety of humanities collections. The pilot project would concentrate on large holdings of papyrological and epigraphic texts from North Africa during the Greek and Roman periods. ($129,828)

A team of scholars from the Digital Archaeological Archive of Comparative Slavery at the Thomas Jefferson Foundation in Virginia, the University of Southampton's Nevis Heritage Project, and the International Slavery Museum in Liverpool is working together on the St. Kitts-Nevis Digital Archaeology Initiative. Together, they plan to develop an integrated digital archive of diverse archaeological and historical data related to the experiences of slaves on sugar plantations in the Caribbean by digitizing and delivering on the Web information from two 18th-century plantations. ($132,832)

The Perseus Digital Library at Tufts University and the Internet Centre at Imperial College London plan to develop Philogrid, a Web resource for scholars of Classical Antiquity. The project will generate a digital collection of fragmentary writings of Greek historians that is designed to interact with multiple source editions; a repository of philological data about the Greco-Roman world; and set of procedures that draws on the recipient’s experience in processing textual materials from Perseus but that can also be extended to other digital collections. ($119,992)