npm package discovery and stats viewer.

Discover Tips

  • General search

    [free text search, go nuts!]

  • Package details

    pkg:[package-name]

  • User packages

    @[username]

Sponsor

Optimize Toolset

I’ve always been into building performant and accessible sites, but lately I’ve been taking it extremely seriously. So much so that I’ve been building a tool to help me optimize and monitor the sites that I build to make sure that I’m making an attempt to offer the best experience to those who visit them. If you’re into performant, accessible and SEO friendly sites, you might like it too! You can check it out at Optimize Toolset.

About

Hi, 👋, I’m Ryan Hefner  and I built this site for me, and you! The goal of this site was to provide an easy way for me to check the stats on my npm packages, both for prioritizing issues and updates, and to give me a little kick in the pants to keep up on stuff.

As I was building it, I realized that I was actually using the tool to build the tool, and figured I might as well put this out there and hopefully others will find it to be a fast and useful way to search and browse npm packages as I have.

If you’re interested in other things I’m working on, follow me on Twitter or check out the open source projects I’ve been publishing on GitHub.

I am also working on a Twitter bot for this site to tweet the most popular, newest, random packages from npm. Please follow that account now and it will start sending out packages soon–ish.

Open Software & Tools

This site wouldn’t be possible without the immense generosity and tireless efforts from the people who make contributions to the world and share their work via open source initiatives. Thank you 🙏

© 2024 – Pkg Stats / Ryan Hefner

bfj

v9.1.1

Published

Big-friendly JSON. Asynchronous streaming functions for large JSON data sets.

Downloads

15,858,690

Readme

BFJ

Build status Package status Downloads License

Big-Friendly JSON. Asynchronous streaming functions for large JSON data sets.

Why would I want those?

If you need to parse huge JSON strings or stringify huge JavaScript data sets, it monopolises the event loop and can lead to out-of-memory exceptions. BFJ implements asynchronous functions and uses pre-allocated fixed-length arrays to try and alleviate those issues.

Is it fast?

No.

BFJ yields frequently to avoid monopolising the event loop, interrupting its own execution to let other event handlers run. The frequency of those yields can be controlled with the yieldRate option, but fundamentally it is not designed for speed.

Furthermore, when serialising data to a stream, BFJ uses a fixed-length buffer to avoid exhausting available memory. Whenever that buffer is full, serialisation is paused until the receiving stream processes some more data, regardless of the value of yieldRate. You can control the size of the buffer using the bufferLength option but really, if you need quick results, BFJ is not for you.

What functions does it implement?

Nine functions are exported.

Five are concerned with parsing, or turning JSON strings into JavaScript data:

  • read asynchronously parses a JSON file from disk.

  • parse and unpipe are for asynchronously parsing streams of JSON.

  • match selectively parses individual items from a JSON stream.

  • walk asynchronously walks a stream, emitting events as it encounters JSON tokens. Analagous to a SAX parser.

The other four functions handle the reverse transformations, serialising JavaScript data to JSON:

  • write asynchronously serialises data to a JSON file on disk.

  • streamify asynchronously serialises data to a stream of JSON.

  • stringify asynchronously serialises data to a JSON string.

  • eventify asynchronously traverses a data structure depth-first, emitting events as it encounters items. By default it coerces promises, buffers and iterables to JSON-friendly values.

How do I install it?

If you're using npm:

npm i bfj --save

Or if you just want the git repo:

git clone [email protected]:philbooth/bfj.git

How do I read a JSON file?

const bfj = require('bfj');

bfj.read(path, options)
  .then(data => {
    // :)
  })
  .catch(error => {
    // :(
  });

read returns a promise and asynchronously parses a JSON file from disk.

It takes two arguments; the path to the JSON file and an options object.

If there are no syntax errors, the returned promise is resolved with the parsed data. If syntax errors occur, the promise is rejected with the first error.

How do I parse a stream of JSON?

const bfj = require('bfj');

// By passing a readable stream to bfj.parse():
bfj.parse(fs.createReadStream(path), options)
  .then(data => {
    // :)
  })
  .catch(error => {
    // :(
  });

// ...or by passing the result from bfj.unpipe() to stream.pipe():
request({ url }).pipe(bfj.unpipe((error, data) => {
  if (error) {
    // :(
  } else {
    // :)
  }
}))
  • parse returns a promise and asynchronously parses a stream of JSON data.

    It takes two arguments; a readable stream from which the JSON will be parsed and an options object.

    If there are no syntax errors, the returned promise is resolved with the parsed data. If syntax errors occur, the promise is rejected with the first error.

  • unpipe returns a writable stream that can be passed to stream.pipe, then parses JSON data read from the stream.

    It takes two arguments; a callback function that will be called after parsing is complete and an options object.

    If there are no errors, the callback is invoked with the result as the second argument. If errors occur, the first error is passed the callback as the first argument.

How do I selectively parse individual items from a JSON stream?

const bfj = require('bfj');

// Call match with your stream and a selector predicate/regex/JSONPath/string
const dataStream = bfj.match(jsonStream, selector, options);

// Get data out of the returned stream with event handlers
dataStream.on('data', item => { /* ... */ });
dataStream.on('end', () => { /* ... */);
dataStream.on('error', () => { /* ... */);
dataStream.on('dataError', () => { /* ... */);

// ...or you can pipe it to another stream
dataStream.pipe(someOtherStream);

match returns a readable, object-mode stream and asynchronously parses individual matching items from an input JSON stream.

It takes three arguments: a readable stream from which the JSON will be parsed; a selector argument for determining matches, which may be a string, a regular expression, a JSONPath expression, or a predicate function; and an options object.

If the selector is a string, it will be compared to property keys to determine whether each item in the data is a match. If it is a regular expression, the comparison will be made by calling the RegExp test method with the property key. If it is a JSONPath expression, it must start with $. to identify the root node and only use child scope expressions for subsequent nodes. Predicate functions will be called with three arguments: key, value and depth. If the result of the predicate is a truthy value then the item will be deemed a match.

In addition to the regular options accepted by other parsing functions, you can also specify minDepth to only apply the selector to certain depths. This can improve performance and memory usage, if you know that you're not interested in parsing top-level items.

If there are any syntax errors in the JSON, a dataError event will be emitted. If any other errors occur, an error event will be emitted.

How do I write a JSON file?

const bfj = require('bfj');

bfj.write(path, data, options)
  .then(() => {
    // :)
  })
  .catch(error => {
    // :(
  });

write returns a promise and asynchronously serialises a data structure to a JSON file on disk. The promise is resolved when the file has been written, or rejected with the error if writing failed.

It takes three arguments; the path to the JSON file, the data structure to serialise and an options object.

How do I create a stream of JSON?

const bfj = require('bfj');

const stream = bfj.streamify(data, options);

// Get data out of the stream with event handlers
stream.on('data', chunk => { /* ... */ });
stream.on('end', () => { /* ... */);
stream.on('error', () => { /* ... */);
stream.on('dataError', () => { /* ... */);

// ...or you can pipe it to another stream
stream.pipe(someOtherStream);

streamify returns a readable stream and asynchronously serialises a data structure to JSON, pushing the result to the returned stream.

It takes two arguments; the data structure to serialise and an options object.

If there a circular reference is encountered in the data and options.circular is not set to 'ignore', a dataError event will be emitted. If any other errors occur, an error event will be emitted.

How do I create a JSON string?

const bfj = require('bfj');

bfj.stringify(data, options)
  .then(json => {
    // :)
  })
  .catch(error => {
    // :(
  });

stringify returns a promise and asynchronously serialises a data structure to a JSON string. The promise is resolved to the JSON string when serialisation is complete.

It takes two arguments; the data structure to serialise and an options object.

What other methods are there?

bfj.walk (stream, options)

const bfj = require('bfj');

const emitter = bfj.walk(fs.createReadStream(path), options);

emitter.on(bfj.events.array, () => { /* ... */ });
emitter.on(bfj.events.object, () => { /* ... */ });
emitter.on(bfj.events.property, name => { /* ... */ });
emitter.on(bfj.events.string, value => { /* ... */ });
emitter.on(bfj.events.number, value => { /* ... */ });
emitter.on(bfj.events.literal, value => { /* ... */ });
emitter.on(bfj.events.endArray, () => { /* ... */ });
emitter.on(bfj.events.endObject, () => { /* ... */ });
emitter.on(bfj.events.error, error => { /* ... */ });
emitter.on(bfj.events.dataError, error => { /* ... */ });
emitter.on(bfj.events.end, () => { /* ... */ });

walk returns an event emitter and asynchronously walks a stream of JSON data, emitting events as it encounters tokens.

It takes two arguments; a readable stream from which the JSON will be read and an options object.

The emitted events are defined as public properties of an object, bfj.events:

  • bfj.events.array indicates that an array context has been entered by encountering the [ character.

  • bfj.events.endArray indicates that an array context has been left by encountering the ] character.

  • bfj.events.object indicates that an object context has been entered by encountering the { character.

  • bfj.events.endObject indicates that an object context has been left by encountering the } character.

  • bfj.events.property indicates that a property has been encountered in an object. The listener will be passed the name of the property as its argument and the next event to be emitted will represent the property's value.

  • bfj.events.string indicates that a string has been encountered. The listener will be passed the value as its argument.

  • bfj.events.stringChunk indicates that a string chunk has been encountered if the stringChunkSize option was set. The listener will be passed the chunk as its argument.

  • bfj.events.number indicates that a number has been encountered. The listener will be passed the value as its argument.

  • bfj.events.literal indicates that a JSON literal (either true, false or null) has been encountered. The listener will be passed the value as its argument.

  • bfj.events.error indicates that an error was caught from one of the event handlers in user code. The listener will be passed the Error instance as its argument.

  • bfj.events.dataError indicates that a syntax error was encountered in the incoming JSON stream. The listener will be passed an Error instance decorated with actual, expected, lineNumber and columnNumber properties as its argument.

  • bfj.events.end indicates that the end of the input has been reached and the stream is closed.

  • bfj.events.endLine indicates that a root-level newline character has been encountered in an NDJSON stream. Only emitted if the ndjson option is set.

If you are using bfj.walk to sequentially parse items in an array, you might also be interested in the bfj-collections module.

bfj.eventify (data, options)

const bfj = require('bfj');

const emitter = bfj.eventify(data, options);

emitter.on(bfj.events.array, () => { /* ... */ });
emitter.on(bfj.events.object, () => { /* ... */ });
emitter.on(bfj.events.property, name => { /* ... */ });
emitter.on(bfj.events.string, value => { /* ... */ });
emitter.on(bfj.events.number, value => { /* ... */ });
emitter.on(bfj.events.literal, value => { /* ... */ });
emitter.on(bfj.events.endArray, () => { /* ... */ });
emitter.on(bfj.events.endObject, () => { /* ... */ });
emitter.on(bfj.events.error, error => { /* ... */ });
emitter.on(bfj.events.dataError, error => { /* ... */ });
emitter.on(bfj.events.end, () => { /* ... */ });

eventify returns an event emitter and asynchronously traverses a data structure depth-first, emitting events as it encounters items. By default it coerces promises, buffers and iterables to JSON-friendly values.

It takes two arguments; the data structure to traverse and an options object.

The emitted events are defined as public properties of an object, bfj.events:

  • bfj.events.array indicates that an array has been encountered.

  • bfj.events.endArray indicates that the end of an array has been encountered.

  • bfj.events.object indicates that an object has been encountered.

  • bfj.events.endObject indicates that the end of an object has been encountered.

  • bfj.events.property indicates that a property has been encountered in an object. The listener will be passed the name of the property as its argument and the next event to be emitted will represent the property's value.

  • bfj.events.string indicates that a string has been encountered. The listener will be passed the value as its argument.

  • bfj.events.number indicates that a number has been encountered. The listener will be passed the value as its argument.

  • bfj.events.literal indicates that a JSON literal (either true, false or null) has been encountered. The listener will be passed the value as its argument.

  • bfj.events.error indicates that an error was caught from one of the event handlers in user code. The listener will be passed the Error instance as its argument.

  • bfj.events.dataError indicates that a circular reference was encountered in the data and the circular option was not set to 'ignore'. The listener will be passed an Error instance as its argument.

  • bfj.events.end indicates that the end of the data has been reached and no further events will be emitted.

What options can I specify?

Options for parsing functions

  • options.reviver: Transformation function, invoked depth-first against the parsed data structure. This option is analagous to the reviver parameter for JSON.parse.

  • options.yieldRate: The number of data items to process before yielding to the event loop. Smaller values yield to the event loop more frequently, meaning less time will be consumed by bfj per tick but the overall parsing time will be slower. Larger values yield to the event loop less often, meaning slower tick times but faster overall parsing time. The default value is 1024.

  • options.ndjson: If set to true, newline characters at the root level will be treated as delimiters between discrete chunks of JSON. See NDJSON for more information.

  • options.stringChunkSize: For bfj.walk only, set this to the character count at which you wish to chunk strings. Each chunk will be emitted as a bfj.events.stringChunk event, followed by the regular bfj.events.string event after all chunks are emitted.

  • options.recursive: For bfj.match only, set this to true if you wish to match against recursively JSON-parsed strings.

  • options.numbers: For bfj.match only, set this to true if you wish to match against numbers with a string or regular expression selector argument.

  • options.bufferLength: For bfj.match only, the length of the match buffer. Smaller values use less memory but may result in a slower parse time. The default value is 256.

  • options.highWaterMark: For bfj.match only, set this if you would like to pass a value for the highWaterMark option to the readable stream constructor.

Options for serialisation functions

  • options.space: Indentation string or the number of spaces to indent each nested level by. This option is analagous to the space parameter for JSON.stringify.

  • options.promises: By default, promises are coerced to their resolved value. Set this property to 'ignore' for improved performance if you don't need to coerce promises.

  • options.buffers: By default, buffers are coerced using their toString method. Set this property to 'ignore' for improved performance if you don't need to coerce buffers.

  • options.maps: By default, maps are coerced to plain objects. Set this property to 'ignore' for improved performance if you don't need to coerce maps.

  • options.iterables: By default, other iterables (i.e. not arrays, strings or maps) are coerced to arrays. Set this property to 'ignore' for improved performance if you don't need to coerce iterables.

  • options.circular: By default, circular references will cause the write to fail. Set this property to 'ignore' if you'd prefer to silently skip past circular references in the data.

  • options.bufferLength: The length of the write buffer. Smaller values use less memory but may result in a slower serialisation time. The default value is 256.

  • options.highWaterMark: Set this if you would like to pass a value for the highWaterMark option to the readable stream constructor.

  • options.yieldRate: The number of data items to process before yielding to the event loop. Smaller values yield to the event loop more frequently, meaning less time will be consumed by bfj per tick but the overall serialisation time will be slower. Larger values yield to the event loop less often, meaning slower tick times but faster overall serialisation time. The default value is 1024.

Is it possible to pause parsing or serialisation from calling code?

Yes it is! Both walk and eventify decorate their returned event emitters with a pause method that will prevent any further events being emitted. The pause method itself returns a resume function that you can call to indicate that processing should continue.

For example:

const bfj = require('bfj');
const emitter = bfj.walk(fs.createReadStream(path), options);

// Later, when you want to pause parsing:

const resume = emitter.pause();

// Then when you want to resume:

resume();

Can it break long strings into chunks?

Yes. If you pass the stringChunkSize option to bfj.walk, it will emit a bfj.events.stringChunk event for each chunk of the string. The regular bfj.events.string event will still be emitted after all the chunks.

Can it recursively parse JSON nested inside a JSON string?

Yes. If you pass the recursive option to bfj.match, it will recursively parse any string values that satisfy the selector argument. Note the same selector is applied to every level of recursion, so this works best in combination with selectors that are predicate functions.

Can it handle newline-delimited JSON (NDJSON)?

Yes. If you pass the ndjson option to bfj.walk, bfj.match or bfj.parse, newline characters at the root level will act as delimiters between discrete JSON values:

  • bfj.walk will emit a bfj.events.endLine event each time it encounters a newline character.

  • bfj.match will just ignore the newlines while it continues looking for matching items.

  • bfj.parse will resolve with the first value and pause the underlying stream. If it's called again with the same stream, it will resume processing and resolve with the second value. To parse the entire stream, calls should be made sequentially one-at-a-time until the returned promise resolves to undefined (undefined is not a valid JSON token).

bfj.unpipe and bfj.read will not parse NDJSON.

Can I specify a different promise implementation?

Yes. Just pass the Promise option to any method. You might want to try this if you get any out-of-memory errors when parsing huge files.

Is there a change log?

Yes.

How do I set up the dev environment?

The development environment relies on Node.js, ESLint, Mocha, Chai, Proxyquire and Spooks. Assuming that you already have node and NPM set up, you just need to run npm install to install all of the dependencies as listed in package.json.

You can lint the code with the command npm run lint.

You can run the tests with the command npm test.

What versions of Node.js does it support?

As of version 8.0.0, only Node.js versions 18 or greater are supported.

Between versions 3.0.0 and 6.1.2, only Node.js versions 6 or greater were supported.

Until version 2.1.2, only Node.js versions 4 or greater were supported.

What license is it released under?

MIT.