Exporting DB cluster data to Amazon S3
You can export data from a live Amazon Aurora DB cluster to an Amazon S3 bucket. The export process runs in the background and doesn't affect the performance of your active DB cluster.
By default, all data in the DB cluster is exported. However, you can choose to export specific sets of databases, schemas, or tables.
Amazon Aurora clones the DB cluster, extracts data from the clone, and stores the data in an Amazon S3 bucket. The data is stored in an Apache Parquet format that is compressed and consistent. Individual Parquet files are usually 1–10 MB in size.
The faster performance that you can get with exporting snapshot data for Aurora MySQL version 2 and version 3 doesn't apply to exporting DB cluster data. For more information, see Exporting DB cluster snapshot data to Amazon S3.
You're charged for exporting the entire DB cluster, whether you export all or partial data. For more information, see the Amazon Aurora pricing page
After the data is exported, you can analyze the exported data directly through tools like Amazon Athena or Amazon Redshift Spectrum. For more information on using Athena to read Parquet data, see Parquet SerDe in the Amazon Athena User Guide. For more information on using Redshift Spectrum to read Parquet data, see COPY from columnar data formats in the Amazon Redshift Database Developer Guide.
Feature availability and support varies across specific versions of each database engine and across AWS Regions. For more information on version and Region availability of exporting DB cluster data to S3, see Supported Regions and Aurora DB engines for exporting cluster data to Amazon S3.
You use the following process to export DB cluster data to an Amazon S3 bucket. For more details, see the following sections.
Overview of exporting DB cluster data
-
Identify the DB cluster whose data you want to export.
-
Set up access to the Amazon S3 bucket.
A bucket is a container for Amazon S3 objects or files. To provide the information to access a bucket, take the following steps:
-
Identify the S3 bucket where the DB cluster data is to be exported. The S3 bucket must be in the same AWS Region as the DB cluster. For more information, see Identifying the Amazon S3 bucket for export.
-
Create an AWS Identity and Access Management (IAM) role that grants the DB cluster export task access to the S3 bucket. For more information, see Providing access to an Amazon S3 bucket using an IAM role.
-
-
Create a symmetric encryption AWS KMS key for the server-side encryption. The KMS key is used by the cluster export task to set up AWS KMS server-side encryption when writing the export data to S3.
The KMS key policy must include both the
kms:CreateGrant
andkms:DescribeKey
permissions. For more information on using KMS keys in Amazon Aurora, see AWS KMS key management.If you have a deny statement in your KMS key policy, make sure to explicitly exclude the AWS service principal
export.rds.amazonaws.com
.You can use a KMS key within your AWS account, or you can use a cross-account KMS key. For more information, see Using a cross-account AWS KMS key.
-
Export the DB cluster to Amazon S3 using the console or the
start-export-task
CLI command. For more information, see Creating DB cluster export tasks. -
To access your exported data in the Amazon S3 bucket, see Uploading, downloading, and managing objects in the Amazon Simple Storage Service User Guide.
Learn to set up, export, monitor, cancel, and troubleshoot DB cluster export tasks in the following sections.