There are several "Data at Risk" activities underway in various disciplines and institutions. This page contains pointers and brief descriptions to some of the projects we know about.
- The UK Economic and Social Data Service December 2010 news article Data at risk: challenges ahead for acquisition strategies: The UK Data Archive is reviewing its collection development policies and strategies to respond to the expected cuts in funding to public bodies and the resulting data collections that may be at risk. "It is contacting public bodies due to close - or under threat of closure - to identify 'at-risk' data collections that fall under the Archive's remit."
- The British Library Endangered Archives Programme: An announcement of the Endangered Archives Program website, promotion of the program, and a call for grant proposals from researchers working on identifying collections whose preservation has value. This program focuses on many different kinds of artifacts; perhaps scientific data collections can qualify for support.
- Down in the Data Dumps: Researchers Inventory a World of Information: This Scientific American article interviews Martin Hilbert (UCLA Annenberg School for Communication and Journalism) on his work with Priscila López Pavez (Open University of Catalonia in Santiago, Chile) to quantify the amount of scientific data and information being generated. Their work was published in the February 2011 issue of Science Special Online Collection Dealing with Data. This work is cited here because it is about current scientific data and information inventory work.
- From Dusty Punch Cards, New Insights Into Link Between Cholesterol and Heart Disease: A Berkeley Lab News article describing new science discoveries that used old data on old media, in this case, computer punch cards. This is the kind of story that supports the DARTG objectives and work.
- We Need a Research Data Census: Francine Berman's (RPI) December 2010 CACM Viewpoint article that calls for "a Research Data Census [that] would inform cost-effective planning for stewardship of federally funded, shared cyber-infrastructure in the Digital World.” The DARTG work could inform any such census activity.
- Historical records of high frequency sea level data are important too. Most focus of studying long term trends in sea level data has been on monthly means, with the international repository at the Permanent Service for Mean Sea Level, http://www.psmsl.org/. Dr. David Jay and Dr. Stefan Talke of Portland State University have shown that tidal constituents such as the semi-diurnal lunar, or M2, vary with time (Citation: Jay, D. A. (2009), Evolution of tidal amplitudes in the eastern Pacific Ocean, Geophys. Res. Lett., 36, L04603, doi:10.1029/2008GL036185.). This can have ramifications for increasing coastal inundation when high tides coincide with storm-induced events. Availability of long time series of high frequency (hourly or greater) data from global tide gauges is essential for their research. The Joint Archive for Sea Level, http://ilikai.soest.hawaii.edu/UHSLC/jasl.html, an international repository for hourly tide gauge data, is working with Drs. Jay and Talke in tracking down and acquiring such historic records.
- 21st-Century Data Miners Meet 19th-Century Electrical Cables This short article in the June 2011 issue of IEEE Computer describes a project between Columbia University scientists and Consolidated Edison (ConEd) engineers to determine the usefulness of old, raw ConEd data on low-voltage failures (manhole fires, explosions, etc.) in predicting and preventing future events. The result? Yes, raw historical data can be used to predict the future.