EntityTooLarge error when uploading a 5G file to Amazon S3
If you are using aws cli for the upload, you can use 'aws s3 cp' command so it does not require splitting and multi part upload
aws s3 cp masive-file.ova s3://<your-bucket>/<prefix>/masive-file.ova
The trick usually seems to be figuring out how to tell S3 to do a multipart upload. For copying data from HDFS to S3, this can be done by using the s3n filesystem and specifically enabling multipart uploads with fs.s3n.multipart.uploads.enabled=true
This can be done like:
hdfs dfs -Dfs.s3n.awsAccessKeyId=ACCESS_KEY -Dfs.s3n.awsSecretAccessKey=SUPER_SECRET_KEY -Dfs.s3n.multipart.uploads.enabled=true -cp hdfs:///path/to/source/data s3n://bucket/folder/
And further configuration can be found here: https://hadoop.apache.org/docs/stable/hadoop-aws/tools/hadoop-aws/index.html
The object size is limited to 5 TB. The upload size is still 5 GB, as explained in the manual:
Depending on the size of the data you are uploading, Amazon S3 offers the following options:
Upload objects in a single operation—With a single
PUT
operation you can upload objects up to 5 GB in size.Upload objects in parts—Using the Multipart upload API you can upload large objects, up to 5 TB.
http://docs.aws.amazon.com/AmazonS3/latest/dev/UploadingObjects.html
Once you do a multipart upload, S3 validates and recombines the parts, and you then have a single object in S3, up to 5TB in size, that can be downloaded as a single entitity, with a single HTTP GET
request... but uploading is potentially much faster, even on files smaller than 5GB, since you can upload the parts in parallel and even retry the uploads of any parts that didn't succeed on first attempt.