Authorizing access to EMRFS data in Amazon S3
By default, the EMR role for EC2 determines the permissions for accessing EMRFS data in Amazon S3. The IAM policies that are attached to this role apply regardless of the user or group making the request through EMRFS. The default is EMR_EC2_DefaultRole
. For more information, see Service role for cluster EC2 instances (EC2 instance profile).
Beginning with Amazon EMR release version 5.10.0, you can use a security configuration to specify IAM roles for EMRFS. This allows you to customize permissions for EMRFS requests to Amazon S3 for clusters that have multiple users. You can specify different IAM roles for different users and groups, and for different Amazon S3 bucket locations based on the prefix in Amazon S3. When EMRFS makes a request to Amazon S3 that matches users, groups, or the locations that you specify, the cluster uses the corresponding role that you specify instead of the EMR role for EC2. For more information, see Configure IAM roles for EMRFS requests to Amazon S3.
Alternatively, if your Amazon EMR solution has demands beyond what IAM roles for EMRFS provides, you can define a custom credentials provider class, which allows you to customize access to EMRFS data in Amazon S3.
Creating a custom credentials provider for EMRFS data in Amazon S3
To create a custom credentials provider, you implement the AWSCredentialsProvider and the Hadoop Configurable
For a detailed explanation of this approach, see Securely analyze data from another AWS account with EMRFS
The basic steps are as follows:
To specify a custom credentials provider
Create a custom credentials provider class compiled as a JAR file.
Run a script as a bootstrap action to copy the custom credentials provider JAR file to the
/usr/share/aws/emr/emrfs/auxlib
location on the cluster's master node. For more information about bootstrap actions, see (Optional) Create bootstrap actions to install additional software.-
Customize the
emrfs-site
classification to specify the class that you implement in the JAR file. For more information about specifying configuration objects to customize applications, see Configuring applications in the Amazon EMR Release Guide.The following example demonstrates a
create-cluster
command that launches a Hive cluster with common configuration parameters, and also includes:A bootstrap action that runs the script,
, which is saved tocopy_jar_file.sh
in Amazon S3.amzn-s3-demo-bucket
An
emrfs-site
classification that specifies a custom credentials provider defined in the JAR file asMyCustomCredentialsProvider
Note
Linux line continuation characters (\) are included for readability. They can be removed or used in Linux commands. For Windows, remove them or replace with a caret (^).
aws emr create-cluster --applications Name=Hive \ --bootstrap-actions '[{"Path":"
s3://amzn-s3-demo-bucket/copy_jar_file.sh
","Name":"Custom action"}]' \ --ec2-attributes '{"KeyName":"MyKeyPair
","InstanceProfile":"EMR_EC2_DefaultRole",\ "SubnetId":"subnet-xxxxxxxx","EmrManagedSlaveSecurityGroup":"sg-xxxxxxxx",\ "EmrManagedMasterSecurityGroup":"sg-xxxxxxxx"}' \ --service-role EMR_DefaultRole_V2 --enable-debugging --release-labelemr-7.3.0
\ --log-uri 's3n://amzn-s3-demo-bucket
/' --name 'test-awscredentialsprovider-emrfs' \ --instance-type=m5.xlarge --instance-count 3 \ --configurations '[{"Classification":"emrfs-site",\ "Properties":{"fs.s3.customAWSCredentialsProvider":"MyAWSCredentialsProviderWithUri"},\ "Configurations":[]}]'