The REMAP Project: Record Management and Preservation in Digital Repositories

The REMAP Project at the University of Hull has been funded by JISC investigate how record management and digital preservation functions can be best supported in digital repositories. It utilizes the Fedora system.

Here’s an except from the Project Aims page (I have added the links in this excerpt):

The REMAP project has the following aims:

  • To develop Records Management and Digital Preservation (RMDP) workflow(s) in order to understand how a digital repository can support these activities
  • To embed digital repository interaction within working practices for RMDP purposes
  • To further develop the use of a WSBPEL orchestration tool to work with external Web services, including the PRONOM Web services, to provide appropriate metadata and file information for RMDP
  • To develop and test a notification layer that can interact with the orchestration tool and allow RSS
    syndication to individuals alerting them to RMDP tasks
  • To develop and test an intermediate persistence layer to underpin the notification layer and interact
    with the WSBPEL orchestration tool to allow orchestrated workflows to take place over time
  • To test and validate the use of the enhanced WSBPEL tool with institutional staff involved in RMDP activities

Position Papers from the NSF/JISC Repositories Workshop

Position papers from the NSF/JISC Repositories Workshop are now available.

Here’s an excerpt from the Workshop’s Welcome and Themes page:

Here is some background information. A series of recent studies and reports have highlighted the ever-growing importance for all academic fields of data and information in digital formats. Studies have looked at digital information in science and in the humanities; at the role of data in Cyberinfrastructure; at repositories for large-scale digital libraries; and at the challenges of archiving and preservation of digital information. The goal of this workshop is to unite these separate studies. The NSF and JISC share two principal objectives: to develop a road map for research over the next ten years and what to support in the near term.

Here are the position papers:

RLG DigiNews Changes

In the latest issue of RLG DigiNews, Jim Michalko and Lorcan Dempsey announce significant changes to this journal. RLG DigiNews has been a five-star journal that has been essential reading for digital library and preservation specialists. I’d encourage my readers to voice support for its continued excellence as indicated in the below excerpt from the article:

The issue in front of you is the last of RLG DigiNews in its current form. As RLG continues to shape its combination with OCLC and create the new Programs and Research division, we are rethinking the publication program that will support our new agenda while providing readers and authors with the kind of vehicle that supports the re-invention of cultural institutions in the research, teaching, and learning process. RLG DigiNews will be an important part of this program. Expect to see it back with a renewed editorial direction. There’s much to do and coordinate but we’ve committed both the talent and the resources to make this happen. Watch for your next RLG DigiNews no later than January, 2008.

Thank you for your support. Let those responsible know that you’re looking forward to the future.

Thursday’s OAI5 Presentations

Presentations from Thursday’s sessions of the 5th Workshop on Innovations in Scholarly Communication in Geneva are now available.

Here are a few highlights from this major conference:

  • Business Models for Digital Repositories (PowerPoint): "Those setting up, or planning to set up, a digital repository may be interested to know more about what has gone before them. What is involved, what is the cost, how many people are needed, how have others made the case to their institution, and how do you get anything into it once it is built? I have recently undertaken a study of European repository business models for the DRIVER project and will present an overview of the findings."
  • DRIVER: Building a Sustainable Infrastructure of European Scientific Repositories (PowerPoint): "Ten partners from eight countries have entered into an international partnership, to connect and network as a first step more than 50 physically distributed institutional repositories to one, large-scale, virtual Knowledge Base of European research."
  • On the Golden Road : Open Access Publishing in Particle Physics (RealVideo): "A working party works now to bring together funding agencies, laboratories and libraries into a single consortium, called SCOAP3 (Sponsoring Consortium for Open access Publishing in Particle Physics). This consortium will engage with publishers towards building a sustainable model for open access publishing. In this model, subscription fees from multiple institutions are replaced with contracts with publishers of open access journals where the SCOAP3 consortium is a single financial partner."
  • Open Access Forever—Or Five Years, Whichever Comes First: Progress on Preserving the Digital Scholarly Record (RealVideo): "The current state of the curation and preservation of digital scholarship over its entire lifecycle will be reviewed, and progress on problems of specific interest to scholarly communication will be examined. The difficulty of curating the digital scholarly record and preserving it for future generations has important implications for the movement to make that record more open and accessible to the world, so this a timely topic for those who are interested in the future of scholarly communication."

(You may want to download PowerPoint Viewer 2007 if you don’t have PowerPoint 2007).

OpenLOCKSS Project

Led by the University of Glasgow Library, the new JISC-funded OpenLOCKSS project will preserve selected UK open access publications.

Here’s an excerpt from the project proposal:

Although LOCKSS has initially concentrated on negotiations with society and commercial publishers, there has always been an interest in smaller open-access journals, as evidenced by the LOCKSS Humanities Project1, where twelve major US libraries have collaborated to contact more than fifty predominantly North American open access journal titles, enabling them to be preserved within the LOCKSS system. . . .

At present, much open access content is under threat, and is difficult to preserve for posterity under standard arrangements, at least until the British Library, and the other UK national libraries, are able to take a more proactive and comprehensive stance in preserving websites comprising UK output. Many open access journals are small operations, often dependent on one or two enthusiastic editors, often based in university departments and/or small societies, concerned with producing the next issue, and often with very little interest in or knowledge of preservation considerations. Their long term survival beyond the first few issues can often be in doubt, but their content, where appropriate quality controls have been applied, is worthy of preservation.

LOCKSS is an ideal low-cost mechanism for ensuring preservation, provided that appropriate contacts can be made and plug-in developments completed, and sufficient libraries agree to host content, on the Humanities Project model. . . .

Earlier in 2006, a survey was carried out by the LOCKSS Pilot Project, to discover preferences for commercial/society publishers to approach with a view to participating in LOCKSS, and Content Complete Ltd have been undertaking this work, as well as negotiating with the NESLi2 publishers on their LOCKSS participation. . . .

We propose to consider initially the titles with at least six votes (it may not be appropriate to approach all these titles, for example we shall check that all are currently publishing and confirm that they appear to be of appropriate quality), followed by those with five or four votes. We propose that agreements for LOCKSS participation are concluded with at least twelve titles, with fifteen as a likely upper limit.

Is It Time to Stop Printing Journals?

There has been lively discussion about whether it is time to stop printing journals on Liblicense-l of late (March archive and April archive).

Here’s my take.

There are two aspects to this question: (1) Is the print journal format still required for reading purposes?; and (2) Is the print journal format still required to insure full access to journals given that many e-journals are under licenses (and are not owned by libraries) and digital preservation is still in its infancy?

It appears that the answer to (1) may finally be “no, for many users.” However, this may be contingent to some degree on the fact that many commercial e-journals are composed of article PDF files that allow users to print copies that replicate printed articles.

The answer to (2) is less clear, since continued access is contingent on periodic license negotiations and the changing business practices of publishers. Embargoes, ILL restrictions, incomplete back runs, and similar issues may give libraries pause. Very promising digital preservation efforts, such as LOCKSS and Portico still need to pass the test of time. Few libraries believe that publishers by themselves can be relied on to preserve e-journals (for one thing, publishers go out of business).

However, the reality for many libraries is that they have no choice but to dump print whenever possible for strictly economic reasons: print plus electronic is increasingly unaffordable for a variety of reasons.

Last Call for the International Digital Preservation Systems Survey

The Getty Research Institute is conducting an International Digital Preservation Systems Survey. It should yield interesting results, so help out by filling it out. March 30th is the last day.

Here’s a brief description from Karim Boughida and Sally Hubbard at the Getty:

This survey is intended to provide an overview of digital preservation system (DPS) implementation. DPS is defined here as an assembly of computer hardware, software and policies equivalent to a TDR (trusted digital repository) "whose mission is to provide reliable, long-term access to managed digital resources to its designated community, now, and in the future"[1].

The survey was produced by the Getty Research Institute departments of Digital Resource Management and Library Information Systems, and will be distributed primarily among members of the Digital Library Federation (DLF). Results will be shared at the DLF Spring Forum, April 23-25, 2007 (Pasadena, California, USA), and with all respondents who provide contact information. . . .

[1] RLG. 2002. Trusted Digital Repositories: Attributes and Responsibilities. Mountain View, Calif.: RLG, Inc. http://www.rlg.org/en/pdfs/repositories.pdf.

Library of Congress Digital Preservation Web Site Redesigned

The Library of Congress has rolled out a new version of its Digital Preservation Web site.

Here’s a excerpt from Digital Preservation News: March 2007:

As you may have already noticed, the Web site for the National Digital Information Infrastructure and Preservation Program (NDIIPP) has been completely redesigned, with new content sections and more user-friendly text. The goal of the redesign was to make the subject of digital preservation in general and NDIIPP in particular more accessible to a wider audience.

Mellon Grants to CLIR/DLF

The Andrew W. Mellon Foundation has given grants to both the Council on Library and Information Resources and the Digital Library Federation.

Here’s an excerpt from the CLIR grant press release:

The Council on Library and Information Resources (CLIR) has received a three-year, $2.19 million grant from The Andrew W. Mellon Foundation to support general operations. The award will allow CLIR to launch a range of new initiatives in six program areas: cyberinfrastructure, preservation, the next scholar, the emerging library, leadership, and new models. . . .

The breadth of CLIR’s new agenda is represented in six interrelated program areas:

Cyberinfrastructure defines the base technologies of computation and communication, the software programs, and the data-curation and data-preservation programs needed to manage large-scale multimedia data sets, particularly those pertaining to the digital record of our cultural heritage;

Preservation explores sustainable strategies for preserving all media in a complex technological, policy, and economic environment;

The Next Scholar explores and assesses new methodologies, fields of inquiry, strategies for data gathering and collaboration, and modes of communication that are likely to define the next generation of scholars;

The Emerging Library explores and articulates the changing concept of the library with particular focus on its core functions and the consequences for staffing, research and teaching, and economic modeling;

Leadership investigates and defines the skills and expertise needed to administer, inspire, and inform the next generation; and

New Models extrapolates from an array of CLIR’s findings and other related research how academic organizations, institutions, behaviors, and culture may evolve over the coming decade.

Here’s an excerpt from the DLF grant press release:

The Digital Library Federation (DLF) has received an $816,000 grant from The Andrew W. Mellon Foundation for a project designed to make distributed digital collections easier for scholars to use. The project, DLF Aquifer Development for Interoperability Across Scholarly Repositories: American Social History Online, will implement schemas, data models, and technologies to enable scholars to use digital collections as one in a variety of local environments. . . .

The project will address the difficulty that humanities and social science scholars face in finding and using digital materials located in a variety of environments with a bewildering array of interfaces, access protocols, and usage requirements. DLF Aquifer seeks to provide scholars with consistent access to digital library collections pertaining to nineteenth- and twentieth-century U.S. social history across institutional boundaries. The collections are in a variety of formats and include maps and photographs from the Library of Congress historical collections; sheet music from the Sam DeVincent Collection of American Sheet Music at Indiana University; and an array of regional collections, such as Michigan County Histories from the University of Michigan and Tennessee Documentary History from the University of Tennessee, that will facilitate cross-regional studies when combined.

By integrating American Social History Online into a variety of local environments, the project will bring the library to the scholar and make distributed collections available through locally supported tools. The project will take two years to develop and implement, from April 2007 to March 2009.

PRESERV Project Report on Digital Preservation in Institutional Repositories

The JISC PRESERV (Preservation Eprint Services) project has issued a report titled Laying the Foundations for Repository Preservation Services: Final Report from the PRESERV Project.

Here’s an excerpt from the Executive Summary:

The PRESERV project (2005-2007) investigated long-term preservation for institutional repositories (IRs), by identifying preservation services in conjunction with specialists, such as national libraries and archives, and building support for services into popular repository software, in this case EPrints. . . .

PRESERV was able to work with The National Archives, which has produced PRONOMDROID, the pre-eminent tool for file format identification. Instead of linking PRONOM to individual repositories, we linked it to the widely used Registry of Open Access Repositories (ROAR), through an OAI harvesting service. As a result format profiles can be found for over 200 repositories listed in ROAR, what we call the PRONOM-ROAR service. . . .

The lubricant to ease the movement of data between the components of the services model is metadata, notably preservation metadata, which informs, describes and records a range of activities concerned with preserving specific digital objects. PRESERV identified a rich set of preservation metadata, based on the current standard in this area, PREMIS, and where this metadata could be generated in our model. . . .

The most important changes to EPrints software as a result of the project were the addition of a history module to record changes to an object and actions performed on an object, and application programs to package and disseminate data for delivery to an external service using either the Metadata Encoding and Transmission Standard (METS) or the MPEG-21 Part 2: Digital Item Declaration Language (DIDL). One change to the EPrints deposit interface is the option for authors to select a licence indicating rights for allowable use by service providers or users, and others. . . .

PRESERV has identified a powerful and flexible framework in which a wide range of preservation services from many providers can potentially be intermediated to many repositories by other types of repository services. It is proposed to develop and test this framework in the next phase of the project.

A Long Road Ahead for Digitization

The New York Times published an article today ("History, Digitized (and Abridged)") that examines the progress that has been made in digitization in the US. It doesn’t hold many surprises for those in the know, but it might be useful in orienting non-specialists to some of the challenges involved, especially those who think that everything is online on the Internet.

It also has some interesting tidbits, including a chart that shows the holdings of different types of materials in the National Archives and how many items have been digitized for each type.

It has some current cost data from the Library of Congress quoted below:

At the Library of Congress, for example, despite continuing and ambitious digitization efforts, perhaps only 10 percent of the 132 million objects held will be digitized in the foreseeable future. For one thing, costs are prohibitive. Scanning alone on smaller items ranges from $6 to $9 for a 35-millimeter slide, to $7 to $11 a page for presidential papers, to $12 to $25 for poster-size pieces.

It also discusses the copyright laws that apply to sound materials and their impact on digitization efforts:

When it comes to sound recordings, copyright law can introduce additional complications. Recordings made before 1972 are protected under state rather than federal laws, and under a provision of the 1976 Copyright Act, may be entitled to protection under state law until 2067. Also, an additional copyright restriction often applies to the underlying musical composition.

A study published in 2005 by the Library of Congress and the Council on Library and Information Resources found that some 84 percent of historical sound recordings spanning jazz, blues, gospel, country and classical music in the United States, and made from 1890 to 1964, have become virtually inaccessible.

An interesting, well-written article that’s worth a read.

Source: Hafner, Katie. "History, Digitized (and Abridged)." The New York Times, 11 March 2007, BU YT 1, 8-9.

Trustworthy Repositories Audit & Certification: Criteria and Checklist Published

The Center for Research Libraries and RLG Programs have published the Trustworthy Repositories Audit & Certification: Criteria and Checklist.

Here’s an excerpt from the press release:

In 2003, RLG and the US National Archives and Records Administration created a joint task force to address digital repository certification. The goal of the RLG-NARA Task Force on Digital Repository Certification was to develop criteria to identify digital repositories capable of reliably storing, migrating, and providing access to digital collections. With partial funding from the NARA Electronic Records Archives Program, the international task force produced a set of certification criteria applicable to a range of digital repositories and archives, from academic institutional preservation repositories to large data archives and from national libraries to third-party digital archiving services. . . . .

In 2005, the Andrew W. Mellon Foundation awarded funding to the Center for Research Libraries to further establish the documentation requirements, delineate a process for certification, and establish appropriate methodologies for determining the soundness and sustainability of digital repositories. Under this effort, Robin Dale (RLG Programs) and Bernard F. Reilly (President, Center for Research Libraries) created an audit methodology based largely on the checklist, tested it on several major digital repositories, including the E-Depot at the Koninklijke Bibliotheek in the Netherlands, the Inter-University Consortium for Political and Social Research, and Portico.

Findings and methodologies were shared with those of related working groups in Europe who applied the draft checklist in their own domains: the Digital Curation Center (U.K.), DigitalPreservationEurope (Continental Europe) and NESTOR (Germany). The report incorporates the sum of knowledge and experience, new ideas, techniques, and tools that resulted from cross-fertilization between the U.S. and European efforts. It also includes a discussion of audit and certification criteria and how they can be considered from an organizational perspective.

JISC Report Evaluates CLIR/ARL E-Journal Archiving Report

JISC has just released an evaluation of the 2006 CLIR/ARL report e-Journal Archiving Metes and Bounds: A Survey of the Landscape. The new report, which is by Maggie Jones, is titled Review and Analysis of the CLIR Report E-Journal Archiving Metes and Bounds: A Survey of the Landscape.

Here is an excerpt from the Executive Summary:

Although both legal deposit legislation and institutional repositories are important developments, neither of them can reasonably be expected to provide practical solutions for libraries licensing access to e-journals. In the UK, the archiving clauses in the NESLI licence have provided a measure of security for libraries but in the absence of trusted repositories charged with managing e-journals, these have provided largely theoretical assurance.

There is a pressing requirement for trusted repositories focussed on archiving and preserving e-journals, which are independent of publishers, and which offer services which can safeguard content while sharing costs between libraries and publishers equitably. While the concerns of libraries are much the same as they were when the JISC consultancy on e-journals archiving reported in 2003, there are now a clearer set of options emerging. Over the past few years, a number of promising initiatives have been developed which provide much better prospects for continued access to licensed e-journal content and which offer cost-effective services for libraries and publishers. Twelve of these trusted repositories have been profiled in a recent CLIR survey. Many of them, including Portico, Pub Med Central, CLOCKSS, and LOCKSS are already familiar in the UK.

Despite a rapidly changing landscape, there is nevertheless a powerful momentum, as evidenced in the rapid take-up of two of the services, LOCKSS and Portico. It is also now possible to articulate a set of principles for archiving services, based on practical reality, which can guide decision-making. The CLIR survey provides a valuable catalyst which the forthcoming BL/DPC/JISC E-Journal Archiving and Preservation workshop (27th March 2007) and other mechanisms have the opportunity to take a significant step forward in this crucial area.

Nestor Project Will Continue until 2009

The Nestor (Network of Expertise in Long-Term Storage of Digital Resources) Project will continue operations until 2009.

Here is a brief description of the project from its home page:

The project’s objective is to create a network of expertise in long-term storage of digital resources for Germany. As the perspective of current and future archive users is central to the project, the emphasis is put on long-term accessibility. Within the project the following offers will be created: a web-based information forum, a platform for information and communication, criteria for trusted digital repositories, recommendations for certification procedures of digital repositories, recommendations for collecting guidelines and selection criteria of digital resources to be archived, guidelines and policies, the concept for a permanent organisation form of the network of expertise in digital preservation. The long-term goal is a permanent distributed infrastructure for long-term preservation and long-term accessability of digital resources in Germany comparable e.g. to the Digital Preservation Coalition in the UK.

Two new working groups have been established for phase two of the project: Standards for Metadata, Transfer of Objects to Digital Repositories and Object Access Working Group and Interlinking of eScience and Long-Term Preservation Working Group.

An English version of Nestor’s Criteria Catalogue for Trusted Repositories is now available.

Senate Poised to Slash NDIIPP Funding

The Disruptive Library Technology Jester and Free Range Librarian blogs have sounded a warning that $47 million of unobligated current-year funding for the National Digital Information Infrastructure and Preservation Program is in serious danger of being rescinded.

House Joint Resolution 20 has been passed in the House and is now being considered by the Senate.

The NDIIPP 2005 Annual Review provides a detailed look at the work of this important Library of Congress program.

See Murray’s Jester posting for the cutback details and check out his protest letter to Ohio’s Senators.

International Journal of Digital Curation Launched

The Digital Curation Centre has launched the International Journal of Digital Curation, which will be published twice a year in digital form (articles are PDF files). It is edited by Richard Waller, who also edits Ariadne. It is published by UKOLN at the University of Bath, using Open Journal Systems.

The journal is freely available. Although individual articles in the first volume do not have copyright statements, the Submissions page on the journal Web site has the following copyright statement:

Copyright for articles published in this journal is retained by the authors, with first publication rights granted to the University of Bath. By virtue of their appearance in this open access journal, articles are free to use, with proper attribution, in educational and other non-commercial settings.

The first issue includes "Digital Curation, Copyright, and Academic Research"; "Digital Curation for Science, Digital Libraries, and Individuals"; "Scientific Publication Packages—A Selective Approach to the Communication and Archival of Scientific Output"; and other articles.

Digital Preservation via Emulation at Koninklijke Bibliotheek

In a two-year (2005-2007) joint project with Nationaal Archief of the Netherlands, Koninklijke Bibliotheek is developing an emulation system that will allow digital objects in outmoded formats to be utilized in their original form. Regarding the emulation approach, the Koninklijke Bibliotheek says:

Emulation is difficult, the main reason why it is not applied on a large scale. Developing an emulator is complex and time-consuming, especially because the emulated environment must appear authentic en must function accurately as well. When future users are interested in the contents of a file, migration remains the better option. When it is the authentic look and feel and functionality of a file they are after, emulation is worth the effort. This can be the case for PDF documents or websites. For multimedia applications, emulation is in fact the only suitable permanent access strategy.

J. R. van der en Wijngaarden Hoeven’s paper "Modular Emulation as a Long-Term Preservation Strategy for Digital Objects" provides a overview of the emulation approach.

In a related development, a message to padiforum-l on 11/17/06 by Remco Verdegem of the Nationaal Archief of the Netherlands reported on a recent Emulation Expert Meeting, which issued a statement noting the following advantages of emulation for digital preservation purposes:

  • It preserves and permits access to each digital artifact in its original form and format; it may be the only viable approach to preserving digital artifacts that have significant executable and/or interactive behavior.
  • It can preserve digital artifacts of any form or format by saving the original software environments that were used to render those artifacts. A single emulator can preserve artifacts in a vast range of arbitrary formats without the need to understand those formats, and it can preserve huge corpuses without ever requiring conversion or any other processing of individual artifacts.
  • It enables the future generation of surrogate versions of digital artifacts directly from their original forms, thereby avoiding the cumulative corruption that would result from generating each such future surrogate from the previous one.
  • If all emulators are written to run on a stable, thoroughly-specified "emulation virtual machine" (EVM) platform and that virtual machine can be implemented on any future computer, then all emulators can be run indefinitely.

What’s in Your Digital Asset Catastrophe Plan?

Anything? You likely have a disaster plan that addresses digital asset issues. The potential problem with a disaster plan is that it can be grounded in assumptions of relative normalcy: the building burns down, a tornado hits, a lower-category hurricane strikes. It may assume severe damage within a confined area and an unimpaired ability of federal, state, and local agencies (as well as relief organizations) to respond. It may assume that workers are not at the disaster site, that they are relatively unaffected if they are, or that they can evacuate and return with relative ease and speed. It may assume that your offsite tape storage or "hot" backup site is far enough away to be unaffected.

What it probably doesn’t assume is the complete devastation of your city or town; widespread Internet, phone, power, and water outages that could last weeks or months; improbable multiple disasters across a wide region surrounding you; the inability of officials at all levels of government to adequately respond to a quickly deepening crisis; the lack of truly workable evacuation plans; depleted gas supplies for a hundred miles in all directions; your evacuated workers being scattered across a multiple-state area in motels, hotels, and the houses of friends and relatives after trips or 20 to 30 hours in massive traffic jams; your institution’s administration being relocated to a hotel in another city; evacuees ending up in new disaster zones and needing to evacuate yet again; and the possibility of more local post-catastrophe catastrophes in short order.

Here’s some thoughts. You may need to have your backups and hot sites in a part of the country that is unlikely to be experiencing a simultaneous catastrophe. This will not be reliable or convenient if physical data transportation is involved. Your latest data could end up in a delivery service depot in your city or town when the event happens. Even if this doesn’t occur, how frequently will you ship out those updates? Daily? Weekly? Another frequency?

Obviously, a remote hot site is better than just backups. But, if hot sites were cheap, we’d all have them.

In terms of backups, how software/hardware-specific are your systems? Will you have to rebuild a complex hardware/software environment to create a live system? Will the components that you need be readily available? Will you have the means to acquire, house, and implement them?

Lots of copies do keep stuff safe, but there have to be lots of copies. Here are two key issues: copyright and will (no doubt there are many more).

You may have a treasure trove of locally produced digital materials, but, if they are under normal copyright arrangements, no one can replicate them. It took considerable resources to create your digital materials. It’s a natural tendency to want to protect them so that they are accessible, but still yours alone. The question to ask yourself is what do I want to prevent users from doing, now and in the future, with these materials? The Creative Commons licences offer options that bar commercial and derivative use, but still provide the freedom to replicate licensed data. True, if you allow replication, you will not really be able to have unified use statistics, but, in the final analysis, what’s more important statistics or digital asset survival? If you allow derivative works, you may find others add value to your work in surprising and novel ways that benefit your users.

However, merely making your digital assets available doesn’t mean that anyone will go to the trouble of replicating or enhancing them. That requires will on the part of others, and they are busy with their own projects. Moreover, they assume that your digital materials will remain available, not disappear forever in the blink of an eye.

It strikes me that digital asset catastrophe planning may call for cooperative effort by libraries, IT centers, and other data-intensive nonprofit organizations. Perhaps by working jointly economic and logistical barriers can be overcome and cost-effective solutions can emerge.

FCLA Digital Archive

Since 2003, the Florida Center for Library Automation (FCLA) has been creating an IMLS-grant-funded Digital Archive (DA) to serve Florida’s public universities. The DA project’s goals are to: "1) establish a working and trusted digital archive, 2) identify costs involved in all aspects of archiving, and 3) disseminate tools, procedures and results for the widest national impact."

The DA will "accept submission packages from participating partners, ingest digital documents along with the appropriate metadata, and safely store on-site and off-site copies of the files."

The DA is a "dark" archive:

Our original idea, and the one thing that did not change over time, was the idea of
building a dark archive. By “dark” I mean an archive with no real-time, online access to
the content by anyone except repository staff. Dark archives are out of favor right now
but we had some good reasons for it. We serve ten state universities and each of them
has its own presentation systems and some have their own institutional repository
systems. Some of the libraries use FCLA delivery applications but some use their own
applications. Central Florida uses CONTENTdm, South Florida uses SiteSearch and
Florida State uses DigiTool. At FCLA we donÂ’t have the software to replicate these
access functions and we donÂ’t have any desire to; it would cost a great deal to acquire the
software licenses, and it would take a huge amount of staff to support all these
applications. So the idea of our offering presentation services on top of our stored
repository content wasnÂ’t feasible.

Real-life digital preservation efforts are always worth keeping an eye on, and this one is quite ambitious. You can track their progress through their grant page and their publications and presentations page.

The project’s most recent presentation by Priscilla Caplan ("Preservation Rumination: Digital Preservation and the Unfamiliar Future ") is available from OCLC in both PowerPoint and MP3 formats.