The Artima Developer Community
Sponsored Link

Agile Buzz Forum
Data Preservation

0 replies on 1 page.

Welcome Guest
  Sign In

Go back to the topic listing  Back to Topic List Click to reply to this topic  Reply to this Topic Click to search messages in this forum  Search Forum Click for a threaded view of the topic  Threaded View   
Previous Topic   Next Topic
Flat View: This topic has 0 replies on 1 page
James Robertson

Posts: 29924
Nickname: jarober61
Registered: Jun, 2003

David Buck, Smalltalker at large
Data Preservation Posted: Dec 11, 2009 4:29 AM
Reply to this message Reply

This post originated from an RSS feed registered with Agile Buzz by James Robertson.
Original Post: Data Preservation
Feed Title: Cincom Smalltalk Blog - Smalltalk with Rants
Feed URL: http://www.cincomsmalltalk.com/rssBlog/rssBlogView.xml
Feed Description: James Robertson comments on Cincom Smalltalk, the Smalltalk development community, and IT trends and issues in general.
Latest Agile Buzz Posts
Latest Agile Buzz Posts by James Robertson
Latest Posts From Cincom Smalltalk Blog - Smalltalk with Rants

Advertisement

The predecessor system to the Large Hadron Collider produced a lot of data (for its day) - 100 TB. In 1989, that was a lot to archive, and it now exists on magnetic tapes. You would think that the archiving of that would be fairly trivial - after all, 1 TB drives are now $75. However, storing the raw data isn't really the problem - it's all the "meta" around the data that is:

More difficult to preserve is the software necessary to make sense of the data. "Clearly, data is useless without the associated software to read and analyse it," say Holzner and co.
The problem is that computer skills are changing. While much of the original LEP software was written in Fortran, the emphasis today is on C++. How the right kind of Fortran expertise can be preserved for future generations isn't clear.
Another problem is that much of the high-level software used to analyse the data-- user-specific analysis code and plotting macros--was never stored in a central database. Instead, it was kept in personal directories which are deleted a year after somebody leaves a lab. That is now lost.
So while future researchers will be able to access the raw data, they may never know exactly how it was processed into the form that appears in scientific publications.

Given the human nature aspect to those problems, I exoect to read something very similar about the piles of LHC data 30 years from now. The personal directory thing is especially likely to be an ongoing problem...

Technorati Tags: ,

Read: Data Preservation

Topic: Live Streaming on an iPhone Previous Topic   Next Topic Topic: Let me Listen tro my iPod

Sponsored Links



Google
  Web Artima.com   

Copyright © 1996-2019 Artima, Inc. All Rights Reserved. - Privacy Policy - Terms of Use