Batch copy files from Azure Blob Storage to Amazon S3
Content-Length
already on S3)For large workloads (either number of files or bytes), you should run this from AWS in the same region where your bucket is located. This will minimize cost and offer reliable/fast/cheap uploads to S3. You will be billed per byte by Azure for outbound data transfer.
$ npm install --save azure-blob-to-s3
var toS3 = require('azure-blob-to-s3')
toS3({
azure: {
connection: '',
container: 'my-container'
},
aws: {
region: 'us-west-2',
bucket: 'my-bucket',
prefix: 'my-prefix'
}
})
azure-s3 \
--concurrency 10 \
--azure-connection "..." \
--azure-container my-container \
--aws-bucket my-bucket \
--aws-prefix my-prefix
toS3(options)
-> output
Copys files from Azure Blob Storage to AWS S3.
Required
Type: object
Options for configuring the copy.
Type: number
Default: 100
The maximum number of files to concurrently stream from Azure and into S3. This is the highWaterMark
of the file upload stream.
Type: object
Required
Type: string
Azure Blob Storage connection string.
Required
Type: string
Azure Blob Storage container name.
Type: string
Default: undefined
A page token where the file list operation will begin.
Type: object
MIT © Ben Drucker