Skip to content

Latest commit

 

History

History
26 lines (13 loc) · 898 Bytes

File metadata and controls

26 lines (13 loc) · 898 Bytes

About

This is a simple program to test S3 high-level API using Java SDK and Python

  • As Big Data grows in popularity, it becomes more important to move large data sets to and from Amazon S3. You can improve the speed of uploads by parallelizing them.
  • You can break an individual file into multiple parts and upload those parts in parallel by using TransferManager High API
  • The Multipart upload API enables you to upload large objects in parts. You can use this API to upload new large objects

Start using

#java funcion

require: aws-java-sdk-core-2.43 aws-java-sdk-s3-2.43

1.upload one file of big size as a whole and download using threads 2.upload severl files as a list

#python function

require:The program makes use of the FileChunkIO module, so pip install FileChunkIO if it isn’t already installed

upload one file of big size as a whole and download using threads