Abstract

Managing digital data for long-term archival and disaster recovery is a key component of our collective responsibility in managing digital data and metadata. As more and more data are collected digitally and as the metadata for traditional museum collections becomes both digitized and more comprehensive, the need to ensure that these data are safe and accessible in the long term becomes essential. Unfortunately, disasters do occur and many irreplaceable datasets on biodiversity have been permanently lost. Maintaining a long-term archive and putting in place reliable disaster recovery processes can be prohibitively expensive, both in the cost of hardware and software as well as the costs of personnel to manage and maintain an archival system. Traditionally, storing digital data for the long term and ensuring the data are loss-less, safe and completely recoverable when a disaster occurs has been managed on-premises with a combination of on-site and off-site storage. This requires complex data workflows to ensure that all data are securely and redundantly stored in multiple highly dispersed locations to minimize the threat of data loss due to local or regional disasters. Files are often moved multiple times across operating systems and media types on their way to and from a deep archive, increasing the risk of file integrity issues. With the recent advent of an array of Cloud Services from organizations such as Amazon, Microsoft and Google to more focused offerings from Iron Mountain, Atempo and others, we have a number of options for long term archival of digital data. Deep archive solutions, storage where retrieval expected only in the case of a disaster, are offered by many of these organizations at a rate substantially less than their normal data storage fees. The most basic requirement for an archival system is storing multiple replicates of the data in geographically isolated locations with a mechanism for guaranteeing file integrity, usually using a checksum algorithm. Additional components that are integral to a robust archive include a simple metadata search and reliable retrieval. In this presentation, we’ll discuss the need for long term archive and disaster recovery capabilities, detail the current best practices of data archival systems and review a variety of archival options that have become available with Cloud Services.

Highlights

  • Re Gerbracht J, Kelling S offered by many of these organizations at a rate substantially less than their normal data storage fees

  • Managing digital data for long-term archival and disaster recovery is a key component of our collective responsibility in managing digital data and metadata

  • With the recent advent of an array of Cloud Services from organizations such as Amazon, Microsoft and Google to more focused offerings from Iron Mountain, Atempo and others, we have a number of options for long term archival of digital data

Read more

Summary

Introduction

Re Gerbracht J, Kelling S offered by many of these organizations at a rate substantially less than their normal data storage fees. Best Practices for using Cloud Services for Digital Data Archive and Disaster Recovery Corresponding author: Jeff Gerbracht (jag73@cornell.edu), Steve Kelling (stk2@cornell.edu) Received: 13 Apr 2018 | Published: 22 May 2018 Citation: Gerbracht J, Kelling S (2018) Best Practices for using Cloud Services for Digital Data Archive and Disaster Recovery.

Results
Conclusion

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.