Reputation: 29
Some of my objects in my S3 buckets are encrypted with another KMS key.
Which I don't have access to in that case I want to exclude these files.
I have tried using S3 sync and it has parameter --exclude switch which will exclude those files so S3 sync works but the data size is around 100 TB that needs to be completed within 2 days
I want to know if that option is present in S3P as well.
https://www.genui.com/open-source/s3p-massively-parallel-s3-copying
I tried using S3 sync and it works.
aws s3 sync s3://bucket s3://mybucket --exclude "folder/*".
Upvotes: 0
Views: 442
Reputation: 699
S3P-author here. S3P has a number of ways of selecting what files to process. You can see all the options with npx s3p cp --help
. In particular, I'd suggest either:
--filter "js:({Key}) => !/^folder\//.test(Key)"
--stop-at "folder/"
and once with `--start-after "folder/~" might be faster (nothing special about "~" - it's just the last supported character in the character range)Upvotes: 0
Reputation: 270089
An alternate approach is to use Amazon S3 Batch Operations to transfer the files.
It requires an input Manifest file that lists the objects to copy. If it is a large list of objects, you can generate the manifest file through AWS Inventory and then remove the directories/files that you don't want to copy.
Then, create an S3 Batch Operations job to copy the listed objects.
See: Copying objects using S3 Batch Operations - Amazon Simple Storage Service
Upvotes: 1