multi-fs
v1.0.1
Published
A client for doing FS operations in multiple places in sync.
Downloads
38
Readme
multi-fs
A client for doing FS operations in multiple places in sync.
Current supported targets:
- File system root paths
- Remote unix systems (via ssh)
Planned supported targets:
- Rackspace Cloud Files
- Amazon S3
USAGE
var MultiFS = require("multi-fs")
var client = new MultiFS([
// FS is the default type
"/path/to/some/stuff",
// ssh urls are desugared
"ssh://user@host:path/in/home",
// setting special ssh options requires using
// the full object style, though.
{
type: "ssh",
host: "some-host",
user: "some-user",
identity: "/home/.ssh/id_rsa_some_key",
path: "path/in/home"
},
// you can use a variant of the ssh client that does file
// copies by spawning scp. the options are identical to ssh
{
type: "scp",
host: "some-host",
user: "some-user",
identity: "/home/.ssh/id_rsa_some_key",
path: "path/in/home"
},
"scp://user@host:path/in/home",
])
// Paths are not allowed to traverse up past the parent.
client.readdir("foo/bar", function (er, files) {
if (er) {
console.log("it failed!")
} else {
console.log("results:", files)
}
})
Methods
All methods take cb
as their last argument.
Reading functions:
stat(path, cb)
readdir(path, cb)
readFile(path, [encoding], cb)
md5(path, cb)
Writing functions:
writeFile(path, data, [encoding], cb)
writeFilep(path, data, [encoding], cb)
mkdir(path, cb)
mkdirp(path, cb)
unlink(path, cb)
rmdir(path, cb)
rmr(path, cb)
Results
Callbacks are called with the following arguments:
error
First error encountered. If no errors are encountered, but the data returned is not consistent, then it will return an 'Inconsistent Data' error.result
The result of the operation. If consistent results are not found, then this will be set to null.data
Errors, results, and extra metadata from all hosts.
For all methods except readfile
, it performs the operation on all
targets, and will raise an Inconsistent Data
error if they do not
return matching results.
For readfile
, it will call md5
and compare hashes, and then, if
the results all match, it will read the actual file from the first
client that returned an md5 hash.
Calls to writeFile
are atomic on all clients. It will write to a
temporary file like foo.txt.TMP.cafef00d
and then rename to
foo.txt
when finished. If the write fails, it makes a best effort
attempt to unlink the temporary file.
Stat Objects
Because different systems represent file/directory stats differently,
stat calls return a simple object with only isFile
and iDirectory
boolean members as the first argument. The original stat objects from
the underlying systems are returned in the data
argument.
Streams
I think it'd be great to have createReadStream(p, cb)
and
createWriteStream(p, cb)
methods on the client, especially since all
the targets (fs, ssh2, etc.) support streams already.
However, especially for readable streams, it's not at all clear how to
handle inconsistencies. Right now, readFile
will raise an
Inconsistent Data
error if two hosts return different stuff.
However, with a readable stream, it'd have to be checking each chunk
somehow, and that gets pretty complicated.
Probably that "check multiple streams and make sure they're all producing the same data" thing should be a separate module.
For writable streams, it's a bit easier, since it's just a multiplexing pipe thing, but hasn't been done at this time.