[
https://issues.apache.org/jira/browse/HBASE-29434?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=18024501#comment-18024501
]
Dieter De Paepe commented on HBASE-29434:
-----------------------------------------
This issue has been solved by HBASE-29631 and HBASE-29146. Closing it.
> Incremental backups fail during mergeSplitBulkloads
> ---------------------------------------------------
>
> Key: HBASE-29434
> URL: https://issues.apache.org/jira/browse/HBASE-29434
> Project: HBase
> Issue Type: Bug
> Affects Versions: 3.0.0, 4.0.0-alpha-1, 2.6.2
> Reporter: Dieter De Paepe
> Priority: Critical
>
>
> {code:java}
> 2025-06-30 08:24:48,857 ERROR
> org.apache.hadoop.hbase.backup.impl.IncrementalTableBackupClient: Failed to
> run MapReduceHFileSplitterJob
> org.apache.hadoop.mapred.FileAlreadyExistsException: Output directory
> hdfs://hdfsns/tmp/backup/hbase/.tmp/backup_1751271394481/lily_ngdata/CUSTOMER/data
> already exists
> at
> org.apache.hadoop.mapreduce.lib.output.FileOutputFormat.checkOutputSpecs(FileOutputFormat.java:164)
> at
> org.apache.hadoop.mapreduce.JobSubmitter.checkSpecs(JobSubmitter.java:278)
> at
> org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:142)
> at org.apache.hadoop.mapreduce.Job$11.run(Job.java:1677)
> at org.apache.hadoop.mapreduce.Job$11.run(Job.java:1674)
> at
> java.base/java.security.AccessController.doPrivileged(AccessController.java:714)
> at java.base/javax.security.auth.Subject.doAs(Subject.java:525)
> at
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1953)
> at org.apache.hadoop.mapreduce.Job.submit(Job.java:1674)
> at org.apache.hadoop.mapreduce.Job.waitForCompletion(Job.java:1695)
> at
> org.apache.hadoop.hbase.backup.mapreduce.MapReduceHFileSplitterJob.run(MapReduceHFileSplitterJob.java:171)
> at
> org.apache.hadoop.hbase.backup.impl.IncrementalTableBackupClient.mergeSplitBulkloads(IncrementalTableBackupClient.java:219)
> at
> org.apache.hadoop.hbase.backup.impl.IncrementalTableBackupClient.mergeSplitBulkloads(IncrementalTableBackupClient.java:203)
> at
> org.apache.hadoop.hbase.backup.impl.IncrementalTableBackupClient.handleBulkLoad(IncrementalTableBackupClient.java:174)
> at
> org.apache.hadoop.hbase.backup.impl.IncrementalTableBackupClient.execute(IncrementalTableBackupClient.java:309)
> at
> org.apache.hadoop.hbase.backup.impl.BackupAdminImpl.backupTables(BackupAdminImpl.java:594)
> {code}
> Looking at the code, I think this is due to
> "IncrementalTableBackupClient#mergeSplitBulkloads" calling
> "mergeSplitBulkloads(List<String> files, TableName tn)" for both archived and
> non-archived files. If both are non-empty, they'll end up in the same output
> folder.
>
--
This message was sent by Atlassian Jira
(v8.20.10#820010)