awslabs / amazon-s3-data-replication-hub-plugin

The Amazon S3 Transfer Plugin for Data Transfer Hub(https://github.com/awslabs/data-transfer-hub). Transfer objects from S3(in other partition), Alibaba Cloud OSS, Tencent COS, Qiniu Kodo into Amazon S3.
Apache License 2.0
47 stars 16 forks source link

Support/fix for transfers of more than 1TB size objects #96

Closed elderjoe closed 1 year ago

elderjoe commented 1 year ago

Describe the bug As stated in the title, objects more than 1TB are having an issue with the transfer.

Using Data Transfer Hub and this plugin is the reason why this issue is created.

Use Case AMIs needs to be transferred to AWS China and these AMIs are more than 1TB.

Once these AMIs are transferred, EC2 instances will be created.

To Reproduce

Expected behavior AMI should be transferred to the destination S3 bucket.

Please complete the following information about the solution:

Screenshots Error encountered:

image

Note: Role policy have no issue. S3 objects less than 1TB are successfully transferred. See Additional context.

Additional context AMIs less than 1TB are successfully transferred with the same steps in the To Reproduce.

image

Created cases in AWS however, it looks like no one is aware or have knowledge about the solution therefore resolving the issue may take a while.

elderjoe commented 1 year ago

Closing, this repository is not maintained anymore, I guess.

evalzy commented 1 year ago

The repository is keeping update, for the large file we suggest user split into 50MB to get optimal performance. Team is working on the alternative solution to support larger file transfer.

elderjoe commented 1 year ago

Hi @evalzy ,

I won't take much of your time, as we are looking for a solution in our predicament ASAP;

So, is it safe to say for now that for 1TB or higher is not supported by this plugin?

Also, when you meant, "split into 50mb", do you mean split the ami-123abc.bin to 50mb? That would be a lot for 1TB or more.

Looking forward to your reply.

Thanks.

evalzy commented 1 year ago

HI @elderjoe

If a single file reaches the GB level, the transmission failure rate increases. The larger the file, the higher the error rate.

We have the tools provide by S3 can be used for splitting, will back to you soon.

elderjoe commented 1 year ago

Hi @evalzy ,

Thank you.

evalzy commented 1 year ago

HI @elderjoe

You can refer below 2 method to split the big file into small piece then use the DTH to perform transfer task.

Refer 1 : https://medium.com/swlh/automatic-s3-file-splitter-620d04b6e81c

Refer 2: https://stackoverflow.com/questions/57351436/how-to-split-the-big-s3-file-into-multiple-files-based-on-size

elderjoe commented 1 year ago

Hello @evalzy,

Thank you. We will check these and see how it goes.

Really appreciated your help.