jac-s3-sync-aws
v2.0.1
Published
A streaming upload tool for Amazon S3
Downloads
10
Readme
jac-s3-sync-aws
Fork from s3-sync-aws(https://github.com/andreialecu/s3-sync-aws)
The difference between this fork is i add a option dest
.
Installation
npm install jac-s3-sync-aws
Usage
require('jac-s3-sync-aws').createStream([db, ]options)
Creates an upload stream. Passes its options to aws-sdk, so at a minimum you'll need:
key
oraccessKeyId
: Your AWS access key.secret
orsecretAccessKey
: Your AWS secret.bucket
: The bucket to upload to.region
: The region the bucket is in.
The following are also specific to jac-s3-sync-aws:
dest
: Upload folder.concurrency
: The maximum amount of files to upload concurrently.retries
: The maximum number of times to retry uploading a file before failing. By default the value is 7.headers
: Additional parameters for each file, see S3 docs.hashKey
: By default, file hashes are stored based on the file's absolute path. This doesn't work very nicely with temporary files, so you can pass this function in to map the file object to a string key for the hash.acl
: Use a custom ACL header. Defaults topublic-read
.force
: Force s3-sync-aws to overwrite any existing files. Not generally required, since we store a hash and compare it to detect updated files.
You can also store your local cache in S3, provided you pass the following
options, and use getCache
and putCache
(see below) before/after uploading:
cacheDest
: the path to upload your cache backup to in S3.cacheSrc
: the local, temporary, text file to stream to before uploading to S3.
If you want more control over the files and their locations that you're uploading, you can write file objects directly to the stream, e.g.:
var stream = s3sync({
key: process.env.AWS_ACCESS_KEY
, secret: process.env.AWS_SECRET_KEY
, bucket: 'sync-testing'
})
stream.write({
src: __filename
, dest: '/uploader.js'
})
stream.end({
src: __dirname + '/README.md'
, dest: '/README.md'
})
Where src
is the absolute local file path, and dest
is the location to
upload the file to on the S3 bucket.
db
is an optional argument - pass it a level database and it'll keep a
local cache of file hashes, keeping S3 requests to a minimum.
stream.putCache(callback)
Uploads your level cache, if available, to the S3 bucket. This means that your cache only needs to be populated once.
stream.getCache(callback)
Streams a previously uploaded cache from S3 to your local level database.
stream.on('fail', callback)
Emitted when a file has failed to upload. This will be called each time the file is attempted to be uploaded.
Example
Here's an example using level
and readdirp
to upload a local directory to
an S3 bucket:
var level = require('level')
, s3sync = require('jac-s3-sync-aws')
, readdirp = require('readdirp')
// To cache the S3 HEAD results and speed up the
// upload process. Usage is optional.
var db = level(__dirname + '/cache')
var files = readdirp({
root: __dirname
, directoryFilter: ['!.git', '!cache']
})
// Takes the same options arguments as `aws-sdk`,
// plus some additional options listed above
var uploader = s3sync(db, {
key: process.env.AWS_ACCESS_KEY
, secret: process.env.AWS_SECRET_KEY
, bucket: 'sync-testing'
, concurrency: 16
, dest: 'target_folder'
, prefix : 'mysubfolder/' //optional prefix to files on S3
}).on('data', function(file) {
console.log(file.fullPath + ' -> ' + file.url)
})
files.pipe(uploader)
You can find another example which includes remote cache storage at example.js.