How to Design a Modern Data Protection Architecture

Data Protection is changing. Today backup software can recover in-place, instantiating volumes directly on the backup storage which enables organizations to significantly reduce recovery windows. Consequently, the performance of the backup storage hardware matters more than ever. At the same time backup software can scale to manage petabytes of stored backup data, so keeping the hardware costs associated with storing this data in check is also critical. Finally, new compliance regulations mean that data within backups must meet both retention requirements and new data privacy demands.

Software applications like Veeam’s Availability Platform provide the capabilities for organizations to meet all of these requirements. The challenge for organizations is to update the hardware components within their data protection architecture so that the software can deliver fully on the software’s promise.

The Components of a Modern Data Protection Infrastructure

Tier 0 – Snapshot Based Instant Recovery

Snapshots have been available as a form of recovery for over a decade. The problem is that snapshots are unorganized and not typically integrated into the data protection process. As a result of the lack of integration, most organizations only keep a minimal number of snapshots. Storage system snapshots need to be integrated with and managed by the data protection software so that IT administrators have a single point of truth when trying to determine which is the most appropriate data set to recover.

Once snapshots are integrated and managed by the backup solution, the backup solution can combine the snapshot data set with the recovery-in-place feature to provide instant recovery. More critical is that the recovery occurs on the same storage system as the original data set, so it is correctly optimized for performance and runs on the same production network. The result should be a seamless application performance experience.

Several storage systems on the market can retain almost unlimited snapshot copies. The challenge is finding the right data within all those snapshots. Integrating the snapshots with the backup solution enables searching of those snapshots via the backup solution’s search capabilities.

The downside of using snapshots as a data protection solution is that they are dependent on the primary storage system and its volume to remain viable. If the storage system fails, then both the primary volume and its snapshots are also unavailable. For these reasons, IT administrators should use snapshots alongside other data protection tiers.

Tier 1 – Recovery-In-Place from All-Flash Backup

The next tier in recovery is new to many IT planners, an all-flash backup tier. For most organizations, this tier can be relatively small in terms of capacity because it only needs to store enough information to recovery the latest copy of an application or dataset. Not all data needs to be sent to this tier. The backup process should only store protected data from workloads that need extremely rapid recovery, and performance similar to production. All other data can directly write to Tier 2, a higher capacity and lower cost but slower performance tier. Also, the storage making up this tier shouldn’t require many features since it is designed to be a temporary production system. The system used for Tier 1 though should be reliable with multiple points of redundancy since these systems could host production workloads for a period until the production system is repaired or replaced.

The backup software should be intelligent enough to automatically copy data from the snapshot to this Tier 1 storage area, which means that IT administrators can perform backups without impacting production applications. In the event of a storage system failure or problem with the actual disk volume, IT administration can redirect the impacted applications to the Tier 1 system and execute an in-place recovery. The Tier 1 system should provide performance as good, if not better, than the primary storage system.

Tier 2 – Long Term Retention from On-premises Object Store

The chances of an organization needing to restore backup data after even a few weeks of it being created become increasingly rare – and the need to perform an in-place recovery becomes even rarer. At the same time, organizations want to retain data in backups for months or even years to meet internal requirements and external compliance and regulations. Recovery of this retained data though is frequent enough that keeping the data on-premises is more practical than recovering all data from the cloud. Cloud storage becomes more expensive over time, and many restores from a cloud tier may incur significant egress fees.

An on-premises network attached storage system or an object storage system are both viable candidates for this long term retention. An on-premises network attached storage (NAS) or object storage system using high capacity hard disk drives can provide the organization with cost-effective long term storage. The backup software can manage moving this data to the high capacity storage tier as it ages.

Tier 3 – Cloud Storage

IT planners can use the cloud storage tier in several ways. The most obvious way is to use cloud storage as a vault for long term data retention and where data is not accessed again unless multiple disasters occur. The infrequency of access enables organizations to use the least expensive cloud storage tier without worrying about excessive egress fees. It becomes the recovery point of last resort in case a site-wide disaster eliminates all the other copies of data.

The other use case for cloud storage, disaster recovery, also involves using cloud computing. In this use case data is pre-positioned in a storage area accessible by cloud computing resources. This area of cloud storage is higher performance than the vault use case, but it only needs to be large enough to support critical applications where rapid recovery is needed even during a site-wide disaster. More than likely it is similar in size to the Tier 1 storage area described above.

NetApp and Veeam – Pulling the Tiers Together

The challenge with any multi-tier architecture is assembling the parts and then managing data movement between the tiers. A multi-tiered data protection architecture like the one described above often requires dealing with multiple storage vendors and data protection software solutions. NetApp and Veeam can deliver all of these capabilities through an integrated hardware and software solution. NetApp has multiple high-performance storage solutions for primary storage that directly integrate with Veeam Backup and Replication. Veeam can execute and manage the NetApp snapshot process as well as instantiate recovery volumes from them. NetApp also has an ideal set of solutions for Tier 1, its E-Series hybrid flash storage platform. For Tier 2, NetApp provides both high capacity network attached storage and its object storage System, StorageGrid. Veeam can move data, through its tiering capabilities, from the snapshot tier to Tier 1 and then subsequently to high capacity NAS or object storage. Veeam can even move or copy data to the cloud for long term data vaulting or cloud disaster recovery.

Conclusion

Users, Application Owners, and Organizations are all demanding much more from the data protection process. These demands come down to recovering faster and retaining data longer. Software solutions like Veeam’s Availability Platform can meet both rapid recovery demands and long term data retention requirements. NetApp is one of the few companies that can provide an entire range of hardware solutions both for primary storage and for the various tiers of the data protection architecture. The combination of NetApp and Veeam make the cost-effectiveness of a multi-tiered data protection architecture a reality by keeping something that could become very complicated, very simple.

Sign up for our Newsletter. Get updates on our latest articles and webinars, plus EXCLUSIVE subscriber only content.

George Crump is the Chief Marketing Officer at VergeIO, the leader in Ultraconverged Infrastructure. Prior to VergeIO he was Chief Product Strategist at StorONE. Before assuming roles with innovative technology vendors, George spent almost 14 years as the founder and lead analyst at Storage Switzerland. In his spare time, he continues to write blogs on Storage Switzerland to educate IT professionals on all aspects of data center storage. He is the primary contributor to Storage Switzerland and is a heavily sought-after public speaker. With over 30 years of experience designing storage solutions for data centers across the US, he has seen the birth of such technologies as RAID, NAS, SAN, Virtualization, Cloud, and Enterprise Flash. Before founding Storage Switzerland, he was CTO at one of the nation's largest storage integrators, where he was in charge of technology testing, integration, and product selection.

Tagged with: , , , , , , , , ,
Posted in Article

Enter your email address to follow this blog and receive notifications of new posts by email.

Join 25,553 other subscribers
Blog Stats
  • 1,906,180 views
%d bloggers like this: