Cloud Computing: DuraSpace Report to Mellon Foundation

The Andrew W. Mellon Foundation has released a progress report from the DuraSpace project, a joint project of the DSpace Foundation and the Fedora Commons. (Thanks to RepositoryMan.)

Here's an excerpt from "DSpace Foundation and Fedora Commons Receive Grant from the Mellon Foundation for DuraSpace" that describes the project:

Over the next six months funding from the planning grant will allow the organizations to jointly specify and design "DuraSpace," a new web-based service that will allow institutions to easily distribute content to multiple storage providers, both "cloud-based" and institution-based. The idea behind DuraSpace is to provide a trusted, value-added service layer to augment the capabilities of generic storage providers by making stored digital content more durable, manageable, accessible and sharable.

“What Cloud Computing Really Means”

Eric Knorr and Galen Gruman provide a concise overview of "cloud computing" in "What Cloud Computing Really Means."

Here's an excerpt:

Cloud computing comes into focus only when you think about what IT always needs: a way to increase capacity or add capabilities on the fly without investing in new infrastructure, training new personnel, or licensing new software. Cloud computing encompasses any subscription-based or pay-per-use service that, in real time over the Internet, extends IT's existing capabilities.

Leslie Carr on Repositories and Cloud Computing

In "The Cloud, the Researcher and the Repository," Leslie Carr discusses repositories and cloud computing, especially the problem of large file deposit.

Here's an excerpt from:

The solution that Tim [Brody] has come up with is to allow the researcher's desktop environment to directly use EPrints as a file system—you can 'mount' the repository as a network drive on your Windows/Mac/Linux desktop using services like WebDAV or FTP. As far as the user is concerned, they can just drag and drop a whole bunch of files from their documents folders, home directories or DVD-ROMs onto the repository disk, and EPrints will automatically deposit them into a new entry or entries. Of course, you can also do the reverse—copy documents from the repository back onto your desktop, open them directly in applications, or attach them to an email.

Cyberinfrastructure Report: e-IRG White Paper 2008

The e-Infrastructure Reflection Group has released the e-IRG White Paper 2008

Here's an excerpt:

The purpose of the e-IRG White Papers is to provide a "snapshot" of technical, policy and organisational developments in the e-Infrastructure domain, influencing the provision, use and broader uptake of the services. Based on this situation analysis, the White Paper outlines the opportunities and challenges stemming from these developments. This analysis culminates in the e-IRG recommendations, providing suggestions for the next steps to maximise benefits from the considerable investments to the electronic infrastructures supporting research activities made in Europe so far.

The selection of topics for this White Paper was based on the recently re-articulated mission of the e-IRG being "to pave the way towards a general-purpose European e-Infrastructure". The topics that at the moment are most important for accomplishing this mission, were chosen after a broad consultation of experts, and presented to e-IRG delegates during the e-IRG workshop in Zurich (April 2008). The topics deemed to be especially relevant at this moment were:

  • Grid and Cloud computing
  • Security: A holistic approach
  • Education and Training
  • Global Collaboration
  • Sustainability of the computing-related e-Infrastructure
  • Remote Instrumentation
  • Virtualisation

What's a Fast Wide Area Network Data Transfer? Now, It's 114 Gigabits per Second

At SuperComputing 2008, an international team headed by California Institute of Technology researchers demonstrated wide area network, multiple-country data transfers that peaked at 114 gigabits per second and sustained a 110 gigabit per second rate.

Read more about it at "High Energy Physics Team Sets New Data-Transfer World Records."

A Web-Based Resource Model for eScience: Object Reuse & Exchange

Carl Lagoze, Herbert Van de Sompel, Michael Nelson, Simeon Warner, Robert Sanderson, and Pete Johnston have deposited "A Web-Based Resource Model for eScience: Object Reuse & Exchange" in arXiv.org.

Here's the abstract:

Work in the Open Archives Initiative-Object Reuse and Exchange (OAI-ORE) focuses on an important aspect of infrastructure for eScience: the specification of the data model and a suite of implementation standards to identify and describe compound objects. These are objects that aggregate multiple sources of content including text, images, data, visualization tools, and the like. These aggregations are an essential product of eScience, and will become increasingly common in the age of data-driven scholarship. The OAI-ORE specifications conform to the core concepts of the Web architecture and the semantic Web, ensuring that applications that use them will integrate well into the general Web environment.

Presentations from the Oxford Institutional and National Services for Research Data Management Workshop

Presentations from the Institutional and National Services for Research Data Management Workshop at the Oxford Said Business School are now available.

Here's a selection:

Presentations from Reinventing Science Librarianship: Models for the Future

Presentations (usually digital audio and PowerPoint slides) about data curation, e-science, virtual organizations and other topics from the ARL/CNI Fall Forum on Reinventing Science Librarianship: Models for the Future are now available.

Speakers included Sayeed Choudhury, Ron Larsen, Liz Lyon, Richard Luce, and others.

Presentations from eResearch Australasia 2008

Presentations from the eResearch Australasia 2008 conference are available.

Here's a brief selection:

E-Science Talking Points for ARL Deans and Directors

The Association of Research Libraries has released E-Science Talking Points for ARL Deans and Directors.

Here's an excerpt:

The term "e-science" is roughly—though not precisely—synonymous with "Cyberinfrastructure;" where the latter term is prevalent in the United States, e-science predominates in the United Kingdom and elsewhere in Europe. Both terms refer to the use of networked computing technologies to enhance collaboration and innovative methods in research. "e-science," however, has a more specific focus on scientific research, whereas Cyberinfrastructure is more inclusive of fields outside the sciences and engineering, and incorporates greater emphasis on supercomputing resources and innovation.

CERN’s Grid: 100,000 Processors at 140 Scientific Institutions

The Worldwide Large Hadron Collider Computing Grid consortium’s grid is ready to process an anticipated 15 million gigabytes per year of data from the collider. It’s composed of 100,000 processors distributed among 140 scientific institutions.

Read more about it at “CERN Officially Unveils Its Grid: 100,000 Processors, 15 Petabytes a Year” and “The Grid Powers Up to Save Lives and Seek the God Particle.”

APSR Releases Investigating Data Management Practices in Australian Universities

The Australian Partnership for Sustainable Repositories has released Investigating Data Management Practices in Australian Universities.

Here an excerpt from the report's Web page:

In late 2007, The University of Queensland undertook a survey of data management practices among the university’s researchers. This was done in response to the increasing realisation that repositories need to include research data, in addition to the research outputs in print form already included, and to provide information which would enhance the support provided for those engaged in eResearch.

The survey was carried out using the Apollo software developed at The Australian National University and adapted by APSR. Two other universities, The University of Melbourne and the Queensland University of Technology, have now replicated the survey among their own communities, while adding some questions of local interest.

The survey covers questions such as the types of digital data being created (spreadsheets, documents, experimental data, images, fieldwork data, etc), the size of the data collection, software used for data analysis, data storage and backup, application of a data management plan, roles and responsibilities around data management, copyright frameworks, usage of high capacity computing, and much more.

CNI Spring 2008 Task Force Meeting Presentations

Presentations and project briefings from the CNI Spring 2008 Task Force Meeting are available. Podcast interviews with a few attendees are also available.

Here's a selection of project briefings:

Digital Research Data Curation: Overview of Issues, Current Activities, and Opportunities for the Cornell University Library

Cornell University Library's Data Working Group has deposited its Digital Research Data Curation: Overview of Issues, Current Activities, and Opportunities for the Cornell University Library report in the eCommons@Cornell repository.

Here's the abstract:

Advances in computational capacity and tools, coupled with the accelerating collection and accumulation of data in many disciplines, are giving rise to new modes of conducting research. Infrastructure to promote and support the curation of digital research data is not yet fully-developed in all research disciplines, scales, and contexts. Organizations of all kinds are examining and staking out their potential roles in the areas of cyberinfrastructure development, data-driven scholarship, and data curation. The purpose of the Cornell University Library's (CUL) Data Working Group (DaWG) is to exchange information about CUL activities related to data curation, to review and exchange information about developments and activities in data curation in general, and to consider and recommend strategic opportunities for CUL to engage in the area of data curation. This white paper aims to fulfill this last element of the DaWG's charge.

EDUCAUSE Publishes Higher Education IT and Cyberinfrastructure: Integrating Technologies for Scholarship

EDUCAUSE has published Higher Education IT and Cyberinfrastructure: Integrating Technologies for Scholarship.

Here's the abstract:

This 2008 ECAR research study explores higher education’s involvement in five areas of research-related information technologies: high-performance computing resources, cyberinfrastructure applications and tools, data storage and management resources, advanced network infrastructure resources, and resources for collaboration within virtual communities. The report, which is based on results of a quantitative survey of 369 U.S. and Canadian colleges and universities and consultation with cyberinfrastructure experts and 12 university executives and technical staff members, discusses who uses, who provides, and who funds cyberinfrastructure resources as well as how important each technology is and will be to research and teaching.

Presentations from the 152nd ARL Membership Meeting

ARL has released presentations from its 152nd membership meeting, which was titled "Institutional Strategies Supporting E-Scholarship & Multidisciplinary Research."

Here's a selection of the presentations:

National Endowment for the Humanities High Performance Computing Initiative Launched

The National Endowment for the Humanities' Office of Digital Humanities has announced the Humanities High Performance Computing Initiative.

Here's an excerpt from the announcement:

As you may have seen in today's Chronicle of Higher Education, the NEH has just announced our new Humanities High Performance Computing initiative—HHPC for short. Our goal is to start a conversation about how high performance computers—supercomputers—can be used for humanities research. We are working with colleagues at the Department of Energy and the National Science Foundation to provide you with information on how high performance/grid computing and data storage might be used for work in the humanities. We are also announcing a new grant competition with DOE to award time and training on their machines. I urge you to check out our HHPC Resources page for more information.

Here's an excerpt from the Humanities High Performance Computing Resource page:

So what do we mean by "HHPC?" Humanities High-Performance Computing (HHPC) refers to the use of high-performance machines for humanities and social science projects. Currently, only a small number of humanities scholars are taking advantage of high-performance computing. But just as the sciences have, over time, begun to tap the enormous potential of HPC, the humanities are beginning to as well. Humanities scholars often deal with large sets of unstructured data. This might take the form of historical newspapers, books, election data, archaeological fragments, audio or video contents, or a host of others. HHPC offers the humanist opportunities to sort through, mine, and better understand and visualize this data.

How to Deal with 10 Petabytes of Data a Year? CERN’s New Grid

CERN's new Large Hadron Collider, which will come online this summer, is expected to generate 10 petabytes of data a year: roughly 1% of the world's entire data output. To deal with this data, CERN is using grid technology with a fiber optic network that links 55,000 servers in 11 global data centers at speeds that are 10,000 times faster than a normal broadband connection.

CERN's GridCafè Web site provides a concise, clear, and easily understood introduction to CERN's grid and grid technology in general.

Read more about it at "Coming Soon: Superfast Internet."

Essays from the Core Functions of the Research Library in the 21st Century Meeting

The Council on Library and Information Resources has released essays from its recent Core Functions of the Research Library in the 21st Century meeting.

Here's an excerpt from the meeting home page that lists the essays:

"The Future of the Library in the Research University," by Paul Courant

"Accelerating Learning and Discovery: Refining the Role of Academic Librarians," by Andrew Dillon

"A New Value Equation Challenge: The Emergence of eResearch and Roles for Research Libraries," by Richard E. Luce

"Co-teaching: The Library and Me," by Stephen G. Nichols

"Groundskeepers to Gatekeepers: How to Change Faculty Perceptions of Librarians and Ensure the Future of the Research Library," by Daphnee Rentfrow

"The Research Library in the 21st Century: Collecting, Preserving, and Making Accessible Resources for Scholarship," by Abby Smith

"The Role of the Library in 21st Century Scholarly Publishing," by Kate Wittenberg

"Leveraging Digital Technologies in Service to Culture and Society: The Role of Libraries as Collaborators," by Lee Zia