[
https://issues.apache.org/jira/browse/HADOOP-15267?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16386962#comment-16386962
]
Anis Elleuch edited comment on HADOOP-15267 at 3/5/18 11:34 PM:
----------------------------------------------------------------
[[email protected]], I added a new patch HDOPP-15267-002.patch which contains
the integration test.
I tested with a AWS S3 bucket (vadmeste-hadoop, us-east-1) using the following
command:
{{mvn test -Dparallel-tests -Dscale -DtestsThreadCount=8
-Dtest=ITestS3AHugeFilesSSECDiskBlocks}}
Please take a look.
was (Author: vadmeste):
[[email protected]] I added a new patch HDOPP-15267-002.patch which contains
the integration test.
I tested with a AWS S3 bucket (vadmeste-hadoop, us-east-1) using the following
command:
{{mvn test -Dparallel-tests -Dscale -DtestsThreadCount=8
-Dtest=ITestS3AHugeFilesSSECDiskBlocks}}
Please take a look.
> S3A multipart upload fails when SSE-C encryption is enabled
> -----------------------------------------------------------
>
> Key: HADOOP-15267
> URL: https://issues.apache.org/jira/browse/HADOOP-15267
> Project: Hadoop Common
> Issue Type: Sub-task
> Components: fs/s3
> Affects Versions: 3.1.0
> Environment: Hadoop 3.1 Snapshot
> Reporter: Anis Elleuch
> Assignee: Anis Elleuch
> Priority: Critical
> Attachments: HADOOP-15267-001.patch, HADOOP-15267-002.patch
>
>
> When I enable SSE-C encryption in Hadoop 3.1 and set fs.s3a.multipart.size
> to 5 Mb, storing data in AWS doesn't work anymore. For example, running the
> following code:
> {code}
> >>> df1 = spark.read.json('/home/user/people.json')
> >>> df1.write.mode("overwrite").json("s3a://testbucket/people.json")
> {code}
> shows the following exception:
> {code:java}
> com.amazonaws.services.s3.model.AmazonS3Exception: The multipart upload
> initiate requested encryption. Subsequent part requests must include the
> appropriate encryption parameters.
> {code}
> After some investigation, I discovered that hadoop-aws doesn't send SSE-C
> headers in Put Object Part as stated in AWS specification:
> [https://docs.aws.amazon.com/AmazonS3/latest/API/mpUploadUploadPart.html]
> {code:java}
> If you requested server-side encryption using a customer-provided encryption
> key in your initiate multipart upload request, you must provide identical
> encryption information in each part upload using the following headers.
> {code}
>
> You can find a patch attached to this issue for a better clarification of the
> problem.
--
This message was sent by Atlassian JIRA
(v7.6.3#76005)
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]