[
https://issues.apache.org/jira/browse/HADOOP-15469?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16479458#comment-16479458
]
Aaron Fabbri commented on HADOOP-15469:
---------------------------------------
And even before change, files could arrive in job output directory during the
commit process. So the window is just larger, right? My interpretation is:
the job driver / app. master is enforcing the "at most once commit" anyways, so
this is more of a sanity check. I think the docs already spell out the
conflict behavior as happening in job setup.
+1 LGTM
I will apply the patch and run tests while i'm in meetings today, but am ok
with you committing this now based on it being a fairly small change. Will
shout if I see any issues.
> S3A directory committer commit job fails if _temporary directory created
> under dest
> -----------------------------------------------------------------------------------
>
> Key: HADOOP-15469
> URL: https://issues.apache.org/jira/browse/HADOOP-15469
> Project: Hadoop Common
> Issue Type: Sub-task
> Components: fs/s3
> Affects Versions: 3.1.0
> Environment: spark test runs
> Reporter: Steve Loughran
> Assignee: Steve Loughran
> Priority: Major
> Attachments: HADOOP-15469-001.patch
>
>
> The directory staging committer fails in commit job if any temporary
> files/dirs have been created. Spark work can create such a dir for placement
> of absolute files.
> This is because commitJob() looks for the dest dir existing, not containing
> non-hidden files.
> As the comment says, "its kind of superfluous". More specifically, it means
> jobs which would commit with the classic committer & overwrite=false will fail
> Proposed fix: remove the check
--
This message was sent by Atlassian JIRA
(v7.6.3#76005)
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]