As the Hybrid cloud became a standard for enterprise IT infrastructure, enterprises consider public cloud storage as a long-term archiving solution. As a result, most Backup applications and storage appliances are now ready to integrate with Azure, AWS storage API. I thought to share some Day2 challenges while deploy, integrate and manage the backup applications with cloud storage options. Commvault is one of the leaders in enterprise backup tools, so a couple of scenarios will be tested in this series of posts using commvault and AWS s3, Glacier. Below picture depicts the LAB architecture. 1) Cloud Storage integration support 2) where we can fit cloud storage in a 3-2-1 strategy for backups 3) Deduplication, Micro pruning options 4) Encryption 5) Object locking and Ransomware protection 6) Cloud Lifecycle policy Support 7) Disaster recovery within the cloud Commvault seems natively supporting most of the cloud storage API without additional license requirements. Integrating library
AWS DataSync is an online data transfer service that simplifies, automates, and accelerates copying large amounts of data between on-premises storage systems and AWS Storage services and between AWS Storage services. For example, DataSync can copy data between Network File System (NFS), Server Message Block (SMB) file servers, self-managed object storage, S3 buckets, EFS file systems, and Amazon FSx. Above diagram depicts the typical architecture of AWS Datasync services. How it works: 1) Data Sync Service: Service in the AWS cloud, which manages and tracks data sync tasks, schedules 2) Data Sync Agent: A Virtual Appliance with computing power to run scheduled copy, uploading capability and maintain metadata (for full and incremental data transfer ) deployed at on-premise or cloud. Advantages: a) Cost-effective solution for Data Sync task ( service charged for per GB transfer in only) b) Best suited for aggressive deployment with zero-touch existing infrastructure. c) Secure tran