Connecting to DynamoDB with Amazon EMR Serverless
In this tutorial, you upload a subset of data from the United States Board on
Geographic Names
Step 1: Upload data to an Amazon S3 bucket
To create an Amazon S3 bucket, follow the instructions in Creating a bucket in
the Amazon Simple Storage Service Console User Guide. Replace references to
with the name of
your newly created bucket. Now your EMR Serverless application is ready to run
jobs.amzn-s3-demo-bucket
-
Download the sample data archive
features.zip
with the following command.wget https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/samples/features.zip
-
Extract the
features.txt
file from the archive and view the first the few lines in the file:unzip features.zip head features.txt
The result should look similar to the following.
1535908|Big Run|Stream|WV|38.6370428|-80.8595469|794 875609|Constable Hook|Cape|NJ|40.657881|-74.0990309|7 1217998|Gooseberry Island|Island|RI|41.4534361|-71.3253284|10 26603|Boone Moore Spring|Spring|AZ|34.0895692|-111.410065|3681 1506738|Missouri Flat|Flat|WA|46.7634987|-117.0346113|2605 1181348|Minnow Run|Stream|PA|40.0820178|-79.3800349|1558 1288759|Hunting Creek|Stream|TN|36.343969|-83.8029682|1024 533060|Big Charles Bayou|Bay|LA|29.6046517|-91.9828654|0 829689|Greenwood Creek|Stream|NE|41.596086|-103.0499296|3671 541692|Button Willow Island|Island|LA|31.9579389|-93.0648847|98
The fields in each line here indicate a unique identifier, name, type of natural feature, state, latitude in degrees, longitude in degrees, and height in feet.
-
Upload your data to Amazon S3
aws s3 cp features.txt s3://
amzn-s3-demo-bucket
/features/
Step 2: Create a Hive table
Use Apache Spark or Hive to create a new Hive table that contains the uploaded data in Amazon S3.
Step 3: Copy data to DynamoDB
Use Spark or Hive to copy data to a new DynamoDB table.
Step 4: Query data from DynamoDB
Use Spark or Hive to query your DynamoDB table.