• Medientyp: E-Artikel
  • Titel: Best Practices for using Cloud Services for Digital Data Archive and Disaster Recovery
  • Beteiligte: Gerbracht, Jeff; Kelling, Steve
  • Erschienen: Pensoft Publishers, 2018
  • Erschienen in: Biodiversity Information Science and Standards, 2 (2018), Seite e25812
  • Sprache: Ohne Angabe
  • DOI: 10.3897/biss.2.25812
  • ISSN: 2535-0897
  • Schlagwörter: General Medicine
  • Entstehung:
  • Anmerkungen:
  • Beschreibung: Managing digital data for long-term archival and disaster recovery is a key component of our collective responsibility in managing digital data and metadata. As more and more data are collected digitally and as the metadata for traditional museum collections becomes both digitized and more comprehensive, the need to ensure that these data are safe and accessible in the long term becomes essential. Unfortunately, disasters do occur and many irreplaceable datasets on biodiversity have been permanently lost. Maintaining a long-term archive and putting in place reliable disaster recovery processes can be prohibitively expensive, both in the cost of hardware and software as well as the costs of personnel to manage and maintain an archival system. Traditionally, storing digital data for the long term and ensuring the data are loss-less, safe and completely recoverable when a disaster occurs has been managed on-premises with a combination of on-site and off-site storage. This requires complex data workflows to ensure that all data are securely and redundantly stored in multiple highly dispersed locations to minimize the threat of data loss due to local or regional disasters. Files are often moved multiple times across operating systems and media types on their way to and from a deep archive, increasing the risk of file integrity issues. With the recent advent of an array of Cloud Services from organizations such as Amazon, Microsoft and Google to more focused offerings from Iron Mountain, Atempo and others, we have a number of options for long term archival of digital data. Deep archive solutions, storage where retrieval expected only in the case of a disaster, are offered by many of these organizations at a rate substantially less than their normal data storage fees. The most basic requirement for an archival system is storing multiple replicates of the data in geographically isolated locations with a mechanism for guaranteeing file integrity, usually using a checksum algorithm. Additional components that are integral to a robust archive include a simple metadata search and reliable retrieval. In this presentation, we’ll discuss the need for long term archive and disaster recovery capabilities, detail the current best practices of data archival systems and review a variety of archival options that have become available with Cloud Services.
  • Zugangsstatus: Freier Zugang