Sun Microsystems Releases Open APIs for the Sun Open Cloud Platform

Sun Microsystems has released Open API's for its Open Cloud Platform.

Here's an excerpt from the press release:

Today at its CommunityOne developer event, Sun Microsystems, Inc. . . . showcased the Sun Open Cloud Platform, the company's open cloud computing infrastructure, powered by industry-leading software technologies from Sun, including Java, MySQL, OpenSolaris and Open Storage. Signaling a massive opportunity to open the world's nascent cloud market, Sun also outlined that a core element of its strategy is to offer public clouds and previewed plans to launch the Sun Cloud, its first public cloud service targeted at developers, student and startups. . . .

As part of the company's commitment to building communities, Sun also announced the release of a core set of Open APIs, unveiled broad partner support for its cloud platform and demonstrated innovative features of the Sun Cloud. Sun is opening its cloud APIs for public review and comment, so that others building public and private clouds can easily design them for compatibility with the Sun Cloud. Sun's Cloud API specifications are published under the Creative Commons license, which essentially allows anyone to use them in any way. Developers will be able to deploy applications to the Sun Cloud immediately, by leveraging pre-packaged VMIs (virtual machine images) of Sun's open source software, eliminating the need to download, install and configure infrastructure software. To participate in the discussion and development of Sun's Cloud APIs, go to sun.com/cloud.

In related news, according to the Wall Street Journal, IBM is negotiating to acquire Sun Microsystems.

Herbert Van de Sompel et al. on “Adding eScience Assets to the Data Web”

Herbert Van de Sompel et al.'s paper on "Adding eScience Assets to the Data Web" is now available on the Linked Data on the Web (LDOW2009) Web site.

Here's an excerpt:

Aggregations of Web resources are increasingly important in scholarship as it adopts new methods that are data-centric, collaborative, and networked-based. The same notion of aggregations of resources is common to the mashed-up, socially networked information environment of Web 2.0. We present a mechanism to identify and describe aggregations of Web resources that has resulted from the Open Archives Initiative-Object Reuse and Exchange (OAI-ORE) project. The OAI-ORE specifications are based on the principles of the Architecture of the World Wide Web, the Semantic Web, and the Linked Data effort. Therefore, their incorporation into the cyberinfrastructure that supports eScholarship will ensure the integration of the products of scholarly research into the Data Web.

Presentations from the 9th International Bielefeld Conference

Presentations from the 9th International Bielefeld Conference are now available.

Here's a few quick selections:

  • Communicating the Results of Research: How Much Does It Cost, and Who Pays?, Michael Jubb (slides) (audio)
  • IR Also Means Institutional Responsibility, Leo Waaijers (slides) (audio)
  • University Investment in the Library: What's the Return?, Carol Tenopir (slides) (audio)

Cloud Computing: DuraSpace Report to Mellon Foundation

The Andrew W. Mellon Foundation has released a progress report from the DuraSpace project, a joint project of the DSpace Foundation and the Fedora Commons. (Thanks to RepositoryMan.)

Here's an excerpt from "DSpace Foundation and Fedora Commons Receive Grant from the Mellon Foundation for DuraSpace" that describes the project:

Over the next six months funding from the planning grant will allow the organizations to jointly specify and design "DuraSpace," a new web-based service that will allow institutions to easily distribute content to multiple storage providers, both "cloud-based" and institution-based. The idea behind DuraSpace is to provide a trusted, value-added service layer to augment the capabilities of generic storage providers by making stored digital content more durable, manageable, accessible and sharable.

“What Cloud Computing Really Means”

Eric Knorr and Galen Gruman provide a concise overview of "cloud computing" in "What Cloud Computing Really Means."

Here's an excerpt:

Cloud computing comes into focus only when you think about what IT always needs: a way to increase capacity or add capabilities on the fly without investing in new infrastructure, training new personnel, or licensing new software. Cloud computing encompasses any subscription-based or pay-per-use service that, in real time over the Internet, extends IT's existing capabilities.

Leslie Carr on Repositories and Cloud Computing

In "The Cloud, the Researcher and the Repository," Leslie Carr discusses repositories and cloud computing, especially the problem of large file deposit.

Here's an excerpt from:

The solution that Tim [Brody] has come up with is to allow the researcher's desktop environment to directly use EPrints as a file system—you can 'mount' the repository as a network drive on your Windows/Mac/Linux desktop using services like WebDAV or FTP. As far as the user is concerned, they can just drag and drop a whole bunch of files from their documents folders, home directories or DVD-ROMs onto the repository disk, and EPrints will automatically deposit them into a new entry or entries. Of course, you can also do the reverse—copy documents from the repository back onto your desktop, open them directly in applications, or attach them to an email.

Cyberinfrastructure Report: e-IRG White Paper 2008

The e-Infrastructure Reflection Group has released the e-IRG White Paper 2008

Here's an excerpt:

The purpose of the e-IRG White Papers is to provide a "snapshot" of technical, policy and organisational developments in the e-Infrastructure domain, influencing the provision, use and broader uptake of the services. Based on this situation analysis, the White Paper outlines the opportunities and challenges stemming from these developments. This analysis culminates in the e-IRG recommendations, providing suggestions for the next steps to maximise benefits from the considerable investments to the electronic infrastructures supporting research activities made in Europe so far.

The selection of topics for this White Paper was based on the recently re-articulated mission of the e-IRG being "to pave the way towards a general-purpose European e-Infrastructure". The topics that at the moment are most important for accomplishing this mission, were chosen after a broad consultation of experts, and presented to e-IRG delegates during the e-IRG workshop in Zurich (April 2008). The topics deemed to be especially relevant at this moment were:

  • Grid and Cloud computing
  • Security: A holistic approach
  • Education and Training
  • Global Collaboration
  • Sustainability of the computing-related e-Infrastructure
  • Remote Instrumentation
  • Virtualisation

What's a Fast Wide Area Network Data Transfer? Now, It's 114 Gigabits per Second

At SuperComputing 2008, an international team headed by California Institute of Technology researchers demonstrated wide area network, multiple-country data transfers that peaked at 114 gigabits per second and sustained a 110 gigabit per second rate.

Read more about it at "High Energy Physics Team Sets New Data-Transfer World Records."

A Web-Based Resource Model for eScience: Object Reuse & Exchange

Carl Lagoze, Herbert Van de Sompel, Michael Nelson, Simeon Warner, Robert Sanderson, and Pete Johnston have deposited "A Web-Based Resource Model for eScience: Object Reuse & Exchange" in arXiv.org.

Here's the abstract:

Work in the Open Archives Initiative-Object Reuse and Exchange (OAI-ORE) focuses on an important aspect of infrastructure for eScience: the specification of the data model and a suite of implementation standards to identify and describe compound objects. These are objects that aggregate multiple sources of content including text, images, data, visualization tools, and the like. These aggregations are an essential product of eScience, and will become increasingly common in the age of data-driven scholarship. The OAI-ORE specifications conform to the core concepts of the Web architecture and the semantic Web, ensuring that applications that use them will integrate well into the general Web environment.

Presentations from the Oxford Institutional and National Services for Research Data Management Workshop

Presentations from the Institutional and National Services for Research Data Management Workshop at the Oxford Said Business School are now available.

Here's a selection:

Presentations from Reinventing Science Librarianship: Models for the Future

Presentations (usually digital audio and PowerPoint slides) about data curation, e-science, virtual organizations and other topics from the ARL/CNI Fall Forum on Reinventing Science Librarianship: Models for the Future are now available.

Speakers included Sayeed Choudhury, Ron Larsen, Liz Lyon, Richard Luce, and others.

Presentations from eResearch Australasia 2008

Presentations from the eResearch Australasia 2008 conference are available.

Here's a brief selection:

E-Science Talking Points for ARL Deans and Directors

The Association of Research Libraries has released E-Science Talking Points for ARL Deans and Directors.

Here's an excerpt:

The term "e-science" is roughly—though not precisely—synonymous with "Cyberinfrastructure;" where the latter term is prevalent in the United States, e-science predominates in the United Kingdom and elsewhere in Europe. Both terms refer to the use of networked computing technologies to enhance collaboration and innovative methods in research. "e-science," however, has a more specific focus on scientific research, whereas Cyberinfrastructure is more inclusive of fields outside the sciences and engineering, and incorporates greater emphasis on supercomputing resources and innovation.

CERN’s Grid: 100,000 Processors at 140 Scientific Institutions

The Worldwide Large Hadron Collider Computing Grid consortium’s grid is ready to process an anticipated 15 million gigabytes per year of data from the collider. It’s composed of 100,000 processors distributed among 140 scientific institutions.

Read more about it at “CERN Officially Unveils Its Grid: 100,000 Processors, 15 Petabytes a Year” and “The Grid Powers Up to Save Lives and Seek the God Particle.”

APSR Releases Investigating Data Management Practices in Australian Universities

The Australian Partnership for Sustainable Repositories has released Investigating Data Management Practices in Australian Universities.

Here an excerpt from the report's Web page:

In late 2007, The University of Queensland undertook a survey of data management practices among the university’s researchers. This was done in response to the increasing realisation that repositories need to include research data, in addition to the research outputs in print form already included, and to provide information which would enhance the support provided for those engaged in eResearch.

The survey was carried out using the Apollo software developed at The Australian National University and adapted by APSR. Two other universities, The University of Melbourne and the Queensland University of Technology, have now replicated the survey among their own communities, while adding some questions of local interest.

The survey covers questions such as the types of digital data being created (spreadsheets, documents, experimental data, images, fieldwork data, etc), the size of the data collection, software used for data analysis, data storage and backup, application of a data management plan, roles and responsibilities around data management, copyright frameworks, usage of high capacity computing, and much more.

CNI Spring 2008 Task Force Meeting Presentations

Presentations and project briefings from the CNI Spring 2008 Task Force Meeting are available. Podcast interviews with a few attendees are also available.

Here's a selection of project briefings:

Digital Research Data Curation: Overview of Issues, Current Activities, and Opportunities for the Cornell University Library

Cornell University Library's Data Working Group has deposited its Digital Research Data Curation: Overview of Issues, Current Activities, and Opportunities for the Cornell University Library report in the eCommons@Cornell repository.

Here's the abstract:

Advances in computational capacity and tools, coupled with the accelerating collection and accumulation of data in many disciplines, are giving rise to new modes of conducting research. Infrastructure to promote and support the curation of digital research data is not yet fully-developed in all research disciplines, scales, and contexts. Organizations of all kinds are examining and staking out their potential roles in the areas of cyberinfrastructure development, data-driven scholarship, and data curation. The purpose of the Cornell University Library's (CUL) Data Working Group (DaWG) is to exchange information about CUL activities related to data curation, to review and exchange information about developments and activities in data curation in general, and to consider and recommend strategic opportunities for CUL to engage in the area of data curation. This white paper aims to fulfill this last element of the DaWG's charge.

EDUCAUSE Publishes Higher Education IT and Cyberinfrastructure: Integrating Technologies for Scholarship

EDUCAUSE has published Higher Education IT and Cyberinfrastructure: Integrating Technologies for Scholarship.

Here's the abstract:

This 2008 ECAR research study explores higher education’s involvement in five areas of research-related information technologies: high-performance computing resources, cyberinfrastructure applications and tools, data storage and management resources, advanced network infrastructure resources, and resources for collaboration within virtual communities. The report, which is based on results of a quantitative survey of 369 U.S. and Canadian colleges and universities and consultation with cyberinfrastructure experts and 12 university executives and technical staff members, discusses who uses, who provides, and who funds cyberinfrastructure resources as well as how important each technology is and will be to research and teaching.

Presentations from the 152nd ARL Membership Meeting

ARL has released presentations from its 152nd membership meeting, which was titled "Institutional Strategies Supporting E-Scholarship & Multidisciplinary Research."

Here's a selection of the presentations:

National Endowment for the Humanities High Performance Computing Initiative Launched

The National Endowment for the Humanities' Office of Digital Humanities has announced the Humanities High Performance Computing Initiative.

Here's an excerpt from the announcement:

As you may have seen in today's Chronicle of Higher Education, the NEH has just announced our new Humanities High Performance Computing initiative—HHPC for short. Our goal is to start a conversation about how high performance computers—supercomputers—can be used for humanities research. We are working with colleagues at the Department of Energy and the National Science Foundation to provide you with information on how high performance/grid computing and data storage might be used for work in the humanities. We are also announcing a new grant competition with DOE to award time and training on their machines. I urge you to check out our HHPC Resources page for more information.

Here's an excerpt from the Humanities High Performance Computing Resource page:

So what do we mean by "HHPC?" Humanities High-Performance Computing (HHPC) refers to the use of high-performance machines for humanities and social science projects. Currently, only a small number of humanities scholars are taking advantage of high-performance computing. But just as the sciences have, over time, begun to tap the enormous potential of HPC, the humanities are beginning to as well. Humanities scholars often deal with large sets of unstructured data. This might take the form of historical newspapers, books, election data, archaeological fragments, audio or video contents, or a host of others. HHPC offers the humanist opportunities to sort through, mine, and better understand and visualize this data.