Details
-
Sub-task
-
Status: Resolved
-
Major
-
Resolution: Fixed
-
3.3.5
Description
iceberg and hbase could benefit from being able to give a list of individual files to delete -files which may be scattered round the bucket for better read peformance.
Add some new optional interface for an object store which allows a caller to submit a list of paths to files to delete, where
the expectation is
- if a path is a file: delete
- if a path is a dir, outcome undefined
For s3 that'd let us build these into DeleteRequest objects, and submit, without any probes first.
Cherrypicking
when cherrypicking, you must include
- followup commit #6854
- https://issues.apache.org/jira/browse/HADOOP-19196
- test fixes HADOOP-19814 and
HADOOP-19188
Attachments
Issue Links
- causes
-
HADOOP-19184 TestStagingCommitter.testJobCommitFailure failing
-
- Resolved
-
-
HADOOP-19188 TestHarFileSystem and TestFilterFileSystem failing after bulk delete API added
-
- Resolved
-
-
HADOOP-19196 Bulk delete api doesn't take the path to delete as the base path
-
- Resolved
-
- is depended upon by
-
HADOOP-17881 S3A DeleteOperation to parallelize POSTing of bulk deletes
-
- Open
-
-
HADOOP-19131 WrappedIO to export modern filesystem/statistics APIs in a reflection friendly form
-
- Resolved
-
- is duplicated by
-
HADOOP-19191 Batch APIs for delete
-
- Resolved
-
- is related to
-
SPARK-49550 Upgrade Hadoop to 3.4.1
-
- Resolved
-
-
HADOOP-18722 Optimise S3A delete objects when multiObjectDelete is disabled
-
- Open
-
-
HADOOP-19254 Implement bulk delete command as hadoop fs command operation
-
- Open
-
-
HADOOP-19203 WrappedIO BulkDelete API to raise IOEs as UncheckedIOExceptions
-
- Resolved
-
- relates to
-
HADOOP-18420 Optimise S3A’s recursive delete to drop successful S3 keys on retry of S3 DeleteObjects
-
- Open
-
- links to