Aws Resources That Should Be Backed Up And How To

As many organizations have discovered first-hand, the consequences of data loss can be downright devastating, often resulting in prolonged downtime, significant damage to credibility, and major financial losses, both direct and indirect. While AWS has been heralded as a safer, more resilient alternative to on-premise computing, organizations must still think about how they can protect their AWS resources against loss by implementing a sound backup strategy.

Selecting AWS Resources for Backup

According to Amazon, AWS resources are all entities that an organization can work with, including EC2 instances, S3 buckets, and CloudFormation stacks. All AWS resources utilize a pay-as-you-go approach for pricing that’s similar to how utility companies charge for natural gas, water, and electricity.

One major advantage of this approach is that organizations pay only for the resources they really consume. However, this can quickly become a disadvantage in the context of data backups. When each and every backup costs a certain amount of money, it’s critical to avoid backing up data that don’t really need to be backed up.

The most important question all organizations should ask themselves when selecting AWS resources for backup is whether they need bare-metal restore capabilities or if the ability to only restore data is sufficient. Organizations that decide to back up only their data may be fine with using  AWS Data Pipeline to routinely move S3 data to Glacier, while other organizations may be fine with something like EC2/RDS snapshots.

Beyond the scope of the backup, it’s also important to select an appropriate backup scheme. For example, the venerable Grandfather-Father-Son (GFS) backup strategy consists of three backup cycles, providing a satisfactory balance between data retention and cost. GFS allows organizations to rest assured, knowing they are protected even against any “acts of God” type of events that might challenge the fantastic durability of AWS.

Recommended AWS Resources for Backup

Listed below are all AWS services (with their corresponding resources) that should be backed up:

Amazon Aurora (Aurora DB Cluster)

Amazon Aurora is a relational database compatible with MySQL and PostgreSQL. It has been designed with the cloud in mind, combining the performance of enterprise databases with the cost-effectiveness of open-source solutions. While fault-tolerant by design, you can manually take a snapshot of the data in your Aurora DB cluster to retain it beyond the set backup retention period.

Amazon DynamoDB (DynamoDB Tables)

Launched in 2012, Amazon DynamoDB is a key-value and NoSQL document database that’s used by some of the largest companies in the world, including Lyft, Airbnb, and Redfin. It comes with a built-in automated on-demand backup, restore, and point-in-time recovery, making it very easy to create full backups of Amazon DynamoDB tables.

Amazon EC2 (EBS Volumes)

One of Amazon’s central services, Elastic Compute Cloud, better known simply as EC2, is a cloud-computing platform that provides secure, resizable compute capacity that can be obtained with minimal friction. EBS volumes should be regularly backed up using Amazon EBS snapshots, and critical applications should be deployed across multiple EC2 Availability Zones.

Amazon Elastic Block Store (EBS Volumes)

Designed for use with Amazon EC2, Amazon Elastic Block Store is a high-performance block storage service that supports a wide range of workloads, including big data analytics engines, enterprise applications, and relational as well as non-relational databases, just to give a few examples. EBS volumes can be distributed across multiple Availability Zones, and it’s also possible to back them up to Amazon S3.

Amazon Elasticsearch (Elasticsearch Clusters)

Built on Apache Lucene and first released in 2010, Amazon Elasticsearch is an open-source, RESTful, distributed search and analytics engine used for anything from business analytics to full-text search to security intelligence to log analytics. One concept that is at the center of Elasticsearch is the Elasticsearch cluster, a collection of nodes that hold data and provide indexing and search capabilities across them. Elasticsearch clusters, including their settings, node information, index settings, and shard allocation, can be backed with Amazon Elasticsearch Service Index Snapshots.

Amazon Redshift (Redshift Cluster)

Powering mission-critical workloads of many Fortune 500 companies, Amazon Redshift has established itself as a premier internet hosting service and data warehouse product thanks to its performance and scale. While reliability is another chief characteristic of Amazon Redshift, it’s still highly recommended to automatically and manually create snapshots of Redshift clusters, which are then stored in Amazon S3.

Amazon Relational Database Service (RDS Database Instances)

Amazon Relational Database Service, or just Amazon RDS, is a distributed relational database service that provides resizable capacity and cost-efficiency while automating many time-consuming administration tasks, such as backups. That said, it’s still recommended to manually take snapshots of RDS database instances and keep them for as long as needed.

Amazon S3 (S3 Buckets)

Designed for 99.999999999% (11 9’s) of durability, Amazon S3 is an object storage service built on the same architecture Amazon uses for its global e-commerce network. Amazon S3 buckets can be effortlessly backed up to Amazon Glacier, which is Amazon’s data archiving and backup service that stands out with its ultra-affordable storage costs.

Amazon WorkSpaces (User Volumes)

Deployed within an Amazon Virtual Private Network (VPC), Amazon Workspaces is a managed, secure cloud desktop service that helps organizations eliminate many administrative tasks associated with the management of Windows and Linux desktops. Individual user volumes are backed up automatically every 12 hours, but it’s best to also enable Amazon WorkDocs Sync on a WorkSpace to allow users to continuously back up a specific folder to Amazon WorkDocs.

AWS Backup Strategies

There are several possible approaches for backing up AWS resources, with Amazon’s own backup service, AWS Backup, leading the way.

Method 1: AWS Backup

AWS Backup is a fully managed backup service that provides a policy-based solution that makes it easy to centralize and automate the backup of data across AWS services. AWS Backup makes it possible to back up and restore EFS file systems, DynamoDB tables, EBS volumes, RDS databases, and Storage Gateway volumes.

Just like other AWS services, AWS Backup utilizes a pay-as-you-go approach for pricing, charging customers on a per-GB basis. For the first backup of an AWS resource, a full copy of your data is saved, but only the changed part of the AWS resource is saved for each incremental backup.  

To back up AWS resources using AWS Backup:

  • Open the AWS Backup console.
  • Create a backup plan.
  • You can build one from scratch, build one based on an existing backup plan, or build one based on the JSON expression of an existing backup plan.
  • Assign resources to the newly created backup plan using tags or listing the resource IDs directly.

Amazon describes the whole process in detail in its Developer Guide.

Method 2: AWS Lambda

Before Amazon rolled out AWS Backup, many AWS users were taking advantage of the backup capabilities of AWS Lambda, an event-driven, serverless computing platform provided by Amazon. Thanks to AWS Lambda, it’s possible to run backup procedures based on trigger events from other AWS services.

A backup procedure can be triggered when someone writes something to an S3 bucket or when a certain amount of time passes since the last backup. Arguably the biggest downside of AWS Lambda is its relatively steep learning curve and the fact that a simple mistake in a backup script can render the entire backup useless and lead to a large bill from Amazon.

Method 3: In-Cloud Backup Solutions

Many providers of backup services have developed in-cloud backup solutions for AWS. Such solutions boast a number of features not supported by AWS Backup, including the ability to backup AWS resources to competing clouds, perform image-based, incremental, and application-aware backups, effectively use storage space with global data deduplication, compression, and exclusion of swap files, and more.

Included among the major providers of in-cloud backup solutions are Veeam, Commvault, Druva, and Acronis. Choosing an independent in-cloud backup solution can prevent vendor lock-in, but the cost of backup will increase.

Aws Resources That Should Be Backed Up And How To

Smart Scheduling at your fingertips

Go from simple to smart, real-time AWS resource scheduling to save cost and increase team productivity.

Learn More
More Posts

You Might Also Like

Product
Introducing the TotalCloud Smart Scheduler
Announcing the launch of the TotalCloud Resource Scheduler! Take complete advantage of AWS's 'pay for what you consume' model by putting a power control for every AWS resource that you use.
January 14, 2020
Sayonee
AWS Use Case Files
Increase EBS Volume Size In Aws
Learn how a simple workflow can auto-remediate and increase the EBS volume size when disk utilization goes beyond 90%
December 2, 2019
Sayonee
AWS Use Case Files
Aws EC2 Instance CPU Utilization Report
Learn how a single workflow can be used to generate a CPU Utilization Report for EC2 Instances and ensure you aren't over or under utilising them.
October 24, 2019
Sayonee
AWS Use Case Files
Aws Lambda Daily Cost Predictor
Learn how a simple workflow can be used to predict daily lambda costs, that can help prevent 'bill-shocks' and optimize your costs better
October 24, 2019
Sayonee
Cloud Automation
Putting Devops On Autopilot With Lego Like Automation
TotalCloud Workflow Editor assists you to do practically anything you want on your AWS infrastructure.
October 24, 2019
Sayonee
AWS Use Case Files
Reporting Security Groups - Tcp Port 22 (Ssh) Public Ip Access
Learn how to use a simple workflow to generate a report of Security Groups with unrestricted Port 22 access from Public IP
October 24, 2019
Sayonee