- 新しい順
- 投票が多い順
- コメントが多い順
Hi Rainesh,
unless you have already tried those options, there are a couple of approaches:
- Multipart upload: https://docs.aws.amazon.com/AmazonS3/latest/userguide/mpuoverview.html
- transfer family: https://docs.aws.amazon.com/transfer/latest/userguide/getting-started.html
hope it helps ;)
You didn't say how you are uploading the file. I wrote a python application that uploads backups from my backup software, which are large, too. I'm uploading a 90 GB file right now. If that's how you are doing this, I can give you more pointers.
There's another pointer that might help, too, that doesn't matter which method you use. It has to do with increasing the amount of retries and you change your config file for that. Here's the documentation for that. Let me know if you need more information on the python method. https://boto3.amazonaws.com/v1/documentation/api/latest/guide/retries.html
Also, since you are uploading large files and they are failing, you may have failed multipart uploads that AWS charges for. I have two articles on how to create a lifecycle rule that deletes these failed uploads. I would set the time period to a day at first, so you don't have to wait long for them to be deleted. You can't see them in your bucket, but you can see them using storage lens. You may want to open a support ticket with AWS if you've been paying for these files. Be prepared to tell them which buckets and how far to go back in time.
I didn't use transfer family because of the restrictions. https://docs.aws.amazon.com/transfer/latest/userguide/limitations-workflow.html