Answering the Big Data Challenge: Protecting Very High Data Volumes and Billions of files
Enterprises face new storage-related challenges which are an integral part of Big Data projects. Today, these projects impact more industries due to the data volume explosions, along with a growing number of files.
Big Data poses new challenges in terms of backup performance requirements, backup windows, overall storage optimization and the ability to retrieve data granularly and flexibly from long term archives. Gartner defines Big Data with three Vs: large Volume, Variety of formats and Velocity.
Atempo-Digital Archive (ADA) addresses the three Vs.
COVER HPC AND BIG DATA STORAGE NEEDS
EXPLODING BACKUP WINDOWS
Large volumes imply that duplicating or replicating data takes time. There is often a bottleneck somewhere. Parallelizing data flow is key. ADA’s data movers can scale up to any infrastructure, matching the data’s throughput and matching the maximum available bandwidth.
Disk, object storage, optical disk, tape, cloud storage media, including combinations of these, can be based on one or more storage technology. As archiving and backup need to cope with both the budget and performance needs, Atempo-Digital Archive offers a broad storage compatibility list with deep integration guaranteeing optimal performance.
FACING PROPRIETARY BACKUP FORMATS
With Atempo-Digital Archive you won’t have to cope with proprietary formats. Only open formats are used. Maintain the original folder structure of your data, with native format on disk and benefit from the widely used TAR and LTFS on tape. Access to files for recovery is direct and easy. Also ideal for a long-term retention, archive and data sharing with third parties.
SPEED & SCALABILITY
ADA is used by our customers to handle extremely large files as well as hundreds of millions of small files which transit between Petabyte storage spaces. Thanks to tight storage integration, Atempo-Digital Archive can leverage snapshots and fastscan technologies with selected vendors. The large volumes of data are transferred by scalable farms of ADA Data Movers, which can each move several GB/s.
To increase performance, simply add a data mover. Scalability is no longer a concern. In addition, all data movements are restricted to authorized storage spaces and are logged.
You just lost something from storage for which ADA had a duplicated copy. This copy is complete with all the folder and sub-folder structure.
Recovery can either be automated or you can browse the storage manually to get the file(s) you need. Or let’s say that you need to retrieve an archived project or folder, use metadata searches or browse and find the relevant version of your assets and restore it to where you need it.
INTEGRATING OBJECT STORAGE
In the past, large volumes were often stored on tape. Now, when it comes to making full and incremental backups of large amounts of data, tape is challenged by object storage.
Object Storage will be able to restore any version of a file, swiftly. Write speed is best served with the right number of ADA Data Movers that will turn it in a powerful and scalable backup solution for Big Data.