[ 
https://issues.apache.org/jira/browse/HADOOP-13058?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15256332#comment-15256332
 ] 

Steve Loughran commented on HADOOP-13058:
-----------------------------------------

{code}
, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 7.281 sec <<< FAILURE! - in 
org.apache.hadoop.fs.s3a.scale.TestS3AInputStreamPerformance
testTimeToOpenAndReadWholeFileByByte(org.apache.hadoop.fs.s3a.scale.TestS3AInputStreamPerformance)
  Time elapsed: 6.973 sec  <<< ERROR!
com.amazonaws.services.s3.model.AmazonS3Exception: Access Denied (Service: 
Amazon S3; Status Code: 403; Error Code: AccessDenied; Request ID: 
8ECDC7355F5EFCCC)
        at 
com.amazonaws.http.AmazonHttpClient.handleErrorResponse(AmazonHttpClient.java:1182)
        at 
com.amazonaws.http.AmazonHttpClient.executeOneRequest(AmazonHttpClient.java:770)
        at 
com.amazonaws.http.AmazonHttpClient.executeHelper(AmazonHttpClient.java:489)
        at 
com.amazonaws.http.AmazonHttpClient.execute(AmazonHttpClient.java:310)
        at 
com.amazonaws.services.s3.AmazonS3Client.invoke(AmazonS3Client.java:3785)
        at 
com.amazonaws.services.s3.AmazonS3Client.abortMultipartUpload(AmazonS3Client.java:2664)
        at 
com.amazonaws.services.s3.transfer.TransferManager.abortMultipartUploads(TransferManager.java:1222)
        at 
org.apache.hadoop.fs.s3a.S3AFileSystem.initMultipartUploads(S3AFileSystem.java:349)
        at 
org.apache.hadoop.fs.s3a.S3AFileSystem.initialize(S3AFileSystem.java:244)
        at 
org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:2786)
        at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:99)
        at 
org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:2823)
        at org.apache.hadoop.fs.FileSystem$Cache.getUnique(FileSystem.java:2811)
        at org.apache.hadoop.fs.FileSystem.newInstance(FileSystem.java:428)
        at 
org.apache.hadoop.fs.s3a.scale.TestS3AInputStreamPerformance.openFS(TestS3AInputStreamPerformance.java:52)
{code}

> S3A FS failes during init against a read-only FS if multipart purge is enabled
> ------------------------------------------------------------------------------
>
>                 Key: HADOOP-13058
>                 URL: https://issues.apache.org/jira/browse/HADOOP-13058
>             Project: Hadoop Common
>          Issue Type: Sub-task
>          Components: fs/s3
>    Affects Versions: 2.8.0
>            Reporter: Steve Loughran
>
> If you try to open a read-only filesystem, and the multipart upload option is 
> set to purge existing uploads, then the FS fails to load with an access 
> denied exception.
> it should catch the exception, downgrade to a debug and await until a file 
> write operation for access exceptions to reject on access rights.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

Reply via email to