npm package discovery and stats viewer.

Discover Tips

  • General search

    [free text search, go nuts!]

  • Package details

    pkg:[package-name]

  • User packages

    @[username]

Sponsor

Optimize Toolset

I’ve always been into building performant and accessible sites, but lately I’ve been taking it extremely seriously. So much so that I’ve been building a tool to help me optimize and monitor the sites that I build to make sure that I’m making an attempt to offer the best experience to those who visit them. If you’re into performant, accessible and SEO friendly sites, you might like it too! You can check it out at Optimize Toolset.

About

Hi, 👋, I’m Ryan Hefner  and I built this site for me, and you! The goal of this site was to provide an easy way for me to check the stats on my npm packages, both for prioritizing issues and updates, and to give me a little kick in the pants to keep up on stuff.

As I was building it, I realized that I was actually using the tool to build the tool, and figured I might as well put this out there and hopefully others will find it to be a fast and useful way to search and browse npm packages as I have.

If you’re interested in other things I’m working on, follow me on Twitter or check out the open source projects I’ve been publishing on GitHub.

I am also working on a Twitter bot for this site to tweet the most popular, newest, random packages from npm. Please follow that account now and it will start sending out packages soon–ish.

Open Software & Tools

This site wouldn’t be possible without the immense generosity and tireless efforts from the people who make contributions to the world and share their work via open source initiatives. Thank you 🙏

© 2024 – Pkg Stats / Ryan Hefner

beamcoder-quiet

v0.3.2

Published

Node.js native bindings to FFmpeg.

Downloads

5

Readme

CircleCI npm version License

Aerostat Beam Coder (quiet version)

This is the same as Aerostat Beam Coder, except it doesn't print out a GPL message when you load the library, because fuck that shit.

Node.js native bindings to FFmpeg with support for asynchronous processing via streams and promises.

The purpose of this package is to facilitate access to the capabilities of FFmpeg - including media muxing, demuxing, encoding, decoding and filtering - from Node.js applications. Rather than using the filesystem and controlling the FFmpeg as an external command line process, the beam coder executes functions of the FFmpeg libav* libraries directly. Work is configured with Javascript objects and jobs execute over data buffers that are shared between Javascript and C. Long running media processing operations are asynchronous, running as promises that execute native code on a separate thread from the main event loop.

Example

View JPEGs in a browser of any key frame of media files in the current folder, e.g. .MP4 media files on a camera memory card, with the following code snippet. For example, to access a key frame around 42.5 seconds from the start of a file called GOPR9502.MP4:

http://localhost:3000/GOPR9502.MP4/42.5

Beam coder uses promises and so the code for the server works well with Javascript's async/await feature and koa contexts:

const beamcoder = require('beamcoder');
const Koa = require('koa');
const app = new Koa();

app.use(async (ctx) => { // Assume HTTP GET with path /<file_name>/<time_in_s>
  let parts = ctx.path.split('/'); // Split the path into filename and time
  if ((parts.length < 3) || (isNaN(+parts[2]))) return; // Ignore favicon etc..
  let dm = await beamcoder.demuxer('file:' + parts[1]); // Probe the file
  await dm.seek({ time: +parts[2] }); // Seek to the closest keyframe to time
  let packet = await dm.read(); // Find the next video packet (assumes stream 0)
  for ( ; packet.stream_index !== 0 ; packet = await dm.read() );
  let dec = beamcoder.decoder({ demuxer: dm, stream_index: 0 }); // Create a decoder
  let decResult = await dec.decode(packet); // Decode the frame
  if (decResult.frames.length === 0) // Frame may be buffered, so flush it out
    decResult = await dec.flush();
  // Filtering could be used to transform the picture here, e.g. scaling
  let enc = beamcoder.encoder({ // Create an encoder for JPEG data
    name : 'mjpeg', // FFmpeg does not have an encoder called 'jpeg'
    width : dec.width,
    height: dec.height,
    pix_fmt: dec.pix_fmt.indexOf('422') >= 0 ? 'yuvj422p' : 'yuvj420p',
    time_base: [1, 1] });
  let jpegResult = await enc.encode(decResult.frames[0]); // Encode the frame
  await enc.flush(); // Tidy the encoder
  ctx.type = 'image/jpeg'; // Set the Content-Type of the data
  ctx.body = jpegResult.packets[0].data; // Return the JPEG image data
});

app.listen(3000); // Start the server on port 3000

Scope

The developers created beam coder to enable development of highly-scalable frame-by-frame, packet-by-packet, web-fit clustered microservices and lambda functions. Beam coder enables the combination of media IO and processing functions with the comprehensive library of scalable IO modules for Node.js, such as express, koa, ioredis, AWS SDK for Javascript etc..

If you are looking to write your own frame-by-frame transcoder, media mangler or muxer, you are in the right place. However, if you want to control FFmpeg as a command line application over complete files or piped streams from a Node.js application, many other projects are available, such as fluent-ffmpeg.

Does beam coder support X, Y or Z protocol / format / codec / file type / stream type / hardware etc.? If FFmpeg supports it, its possible and likely. Beam coder is self-describing at runtime, allowing exploration of what should be possible. However, you have to start somewhere, and the developers have been testing with the codecs and formats they are familiar with. Issues are expected so please raise any problems or requests for additional features as git hub issues. Even better, raise a pull request to fix a problem or add in missing features. Automated testing will be extended in due course, memory management improved, more examples added and hardware accelerated codecs added.

Beam coder will be a cross-platform module for Windows, Mac and Linux. In this early release, only Windows and Linux are available. The Mac platform will follow shortly. The release version of FFmpeg that beam coder links with is currently 4.1.

Aerostat

Beam coder is the first release of Streampunk Media's Aerostat open-source product set, whereby a conceptual fleet of media-oriented aerostats (blimps, air ships, zeppelins etc.) are launched into the clouds. Media content is beamed between the fleet as if light beams, and beamed to and from locations on the planet surface as required. See also the Aerostat Beam Engine.

Installation

NPM

Pre-requisites

Native packages require a build stage that needs some setup. This includes Python 2.7 (not 3) and build tools.

  1. Install the LTS version of Node.js for your platform, currently the latest v10.
  2. Enable node-gyp - the Node.js native addon build tool for your platform by following the installation instructions.

Note: For MacOSX Mojave, install the following package after xcode-select --install:

/Library/Developer/CommandLineTools/Packages/macOS_SDK_headers_for_macOS_10.14.pkg

This release of Beam coder targets 64-bit (x86_64) architectures.

Installing

Windows

Beam coder is intended to be used as a module/package from other Node.js packages and applications. Install beam coder from the root folder of your package as follows:

npm install beamcoder

This will install all necessary dependencies, download the FFmpeg shared and dev packages with include and library files, and then compile the native extensions.

Note that if you want to use a local version of FFmpeg then, before the install, symbolic link or copy appropriate folders to:

./node_modules/beamcoder/ffmpeg/ffmpeg-4.1-win64-shared
./node_modules/beamcoder/ffmpeg/ffmpeg-4.1-win64-dev

To ensure that sufficient threads are available to process several requests in parallel, set the UV_THREADPOOL_SIZE environment variable, e.g.:

set UV_THREADPOOL_SIZE=32

Linux

On Linux, use the appropriate package manager to install the FFmpeg 4.1 development dependencies first. An error will be printed if these cannot be found at expected locations. For example, on Ubuntu:

sudo add-apt-repository ppa:jonathonf/ffmpeg-4
sudo apt-get update
sudo apt-get install libavcodec-dev libavformat-dev libavdevice-dev libavfilter-dev libavutil-dev libpostproc-dev libswresample-dev libswscale-dev

An example for Debian / Redhat / Fedora / CentOs will follow.

Beam coder is intended to be used as a module/package from other applications. Install beam coder from the root folder of your package as follows:

npm install beamcoder

This will check for all the necessary dependencies and then compile the native extensions.

To ensure that sufficient threads are available to process several requests in parallel, set the UV_THREADPOOL_SIZE environment variable, e.g.:

export UV_THREADPOOL_SIZE=32

Mac

Beam coder is intended to be used as a module/package from other applications. Install beam coder from the root folder of your package as follows:

npm install beamcoder

This will install all the necessary dependencies, check and if necessary install FFmpeg and its dependencies via homebrew and then compile the native extensions.

To ensure that sufficient threads are available to process several requests in parallel, set the UV_THREADPOOL_SIZE environment variable, e.g.:

export UV_THREADPOOL_SIZE=32

Usage

Introduction

The basic usage pattern is as follows:

  1. Use promises to create processing stages for an item of media, for example a demuxer and a decoder. One stage can be used to configure the other (see codec parameters).
  2. In a loop, use sequences of promises to move data through the processing stages, working with the data as required. For example, read some frames from a file and send them to a decoder, receiving uncompressed data for each frame. This data may be suitable for display or analysis.
  3. Flush the buffers of each processing stage, close resources and collect any outstanding frames or packets.

Here is a simple example:

const beamcoder = require('beamcoder');

async function run() {
  let demuxer = await beamcoder.demuxer('/path/to/file.mp4'); // Create a demuxer for a file
  let decoder = beamcoder.decoder({ name: 'h264' }); // Codec asserted. Can pass in demuxer.
  let packet = {};
  for ( let x = 0 ; x < 1000 && packet != null ; x++ ) {
    packet = await format.read(); // Read next frame. Note: returns null for EOF
    if (packet && packet.stream_index === 0) { // Check demuxer to find index of video stream
      let frames = await decoder.decode(packet);
      // Do something with the frame data
      console.log(x, frames.total_time); // Optional log of time taken to decode each frame
    }
  }
  let frames = await decoder.flush(); // Must tell the decoder when we are done
  console.log('flush', frames.total_time, frames.length);
}

run();

For mode examples, see the examples folder.

Values

In design, each of the main structures (struct) of FFmpeg have mappings to Javascript objects. Values are automatically marshalled and unmarshalled between native and Javascript representations. These objects contain an internal instance of the native C structure, have a type property providing the type name, and use getters and setters to expose the underlying native properties. Wherever possible, the names of the properties are the same as those in the C library, meaning that FFmpeg's doxygen documentation can and should be used as a reference.

Factory

To create one of the main data types packet, frame, demuxer, muxer, decoder, filterer, or encoder, use the methods of beamcoder with the same name e.g. packet, frame, etc. With no arguments, the value constructed has FFmpeg's default values. To construct a packet:

let pkt = beamcoder.packet();

To configure the object on initialisation, pass it an options object:

let q = beamcoder.packet({ pts: 10, dts: 10, stream_index: 3 });
console.log(q);
{ type: 'Packet',
  pts: 10,
  dts: 10,
  data: null,
  size: 0,
  stream_index: 3,
  /* ... */ }

Factory methods and their associated introspection methods that allow discovery of what are available are:

  • ...demuxer() - find demuxer input formats with ...demuxers().
  • ...decoder() - find decoder codecs with ...decoders().
  • ...filterer() - find details of filters with ...filters().
  • ...encoder() - find encoder codecs with encoders().
  • ...muxer() - find muxer output formats with ...muxers().
  • ...packet() - create a data packet. See creating packets section.
  • ...frame() - construct frames, normally with a format property that is one of the ...pix_fmts() or ...sample_fmts(). See creating frames section.
  • ...codecParameters - create codec parameters for one of the ...codecs(). See codec parameters section below.

Note some special cases to this rule, such as creating a demuxer from a URL or filename. In this case, a single string URL parameter can be passed to the constructor rather than an options object. See the demuxer section for details. To create a format, a generic description of the format independent of a container, as processed by a muxer or demuxer, use ...format().

Reading and modifying

After construction, Javascript properties of an object can be used in a natural way, with dot notation and array-style access:

let q_pts = q.pts; // Current value of pts for q, q_pts= 10
let same = q['pts'] === q.pts; // true ... the current value of pts for q = 10
q.pts = 12; // Set the value pts for q to 12
q.data = Buffer.from('Text data for this packet.'); // Set packet data

This is achieved with getters and setters. Note that not every openly readable (enumerable) property is also writable. This may depend on the current context. For example, some properties are set only by libav* and not by the user. Others can only be updated by libav* during encoding, even when a parameter of the same name could be set by the user when decoding.

The processing work to convert a value from C to Javascript is only done when each separate property is requested. Bear this in mind before being too liberal with, say, console.log() that enumerates through every property of an object. Encoders and decoders have approximately 130 properties!

JSON

The packet, frame, codec parameters, stream and format (container-independent muxer/demuxer) types have mappings to and from JSON. These only include properties that are not currently set to their default values. This is achieved using a native implementation of the toJSON() method, allowing fast creation of JSON representations with JSON.stringify(). In reverse, to parse JSON representations, pass JSON strings to the factory method for these types. For example:

let p = beamcoder.packet({ pts: 654321, stream_index: 3,
  data: Buffer.alloc(65536), duration: 1920});
console.log(JSON.stringify(p, null, 2));
/* {                     
    "type": "Packet",  // Always includes the type name
    "pts": 654321,     // pts is not the default value
    "size": 65536,     // Size is carried as if a "Content-Length: " HTTP header
    "stream_index": 3, // stream_index always included
    "duration": 1920
  } */
console.log(beamcoder.packet(JSON.stringify(p)));
/* { type: 'Packet',
     pts: 654321,
     dts: null,
     data: null,  // Large binary data has to be carried separately
     size: 0,
     stream_index: 3,
     flags: { KEY: false, CORRUPT: false, DISCARD: false, TRUSTED: false, DISPOSABLE: false },
     side_data: null, // Side data values are serialized to JSON arrays
     duration: 1920,
     pos: -1 } */

Note that large binary data buffers are not included in the JSON and need to be managed separately.

Freeing and deleting

Care has been taken to ensure that the reference-counted then garbage collected data structures of Javascript work in tandem with the allocation and free mechanisms of libav*. As such, there is no explicit requirement to free or delete objects. As with any Javascript application, if you hold onto references to objects when they are no longer required, garbage collection is prevented and this may have a detrimental impact on performance. This is particularly the case for frames and packets that hold references to large data buffers.

Type mappings

Property value mappings from C to Javascript and vice versa are as follows:

  • C int, int64_t and other integer types map to and from Javascript's number type. Try to use integer representations in Javascript wherever possible, e.g. use |0 when dividing. (Note that the full ranges of C int64_t and uint64_t are not yet supported by beam coder. This feature is waiting on Javascript BigInt to move from experimental support to full support in Node.) Values set to AV_NOOPTS_VALUE in C (available as beamcoder.AV_NOOPTS_VALUE) convert to null in Javascript.
  • C float and double types map to and from Javascript's number type. Note that as the underlying representation of floating point values is double length in Javascript, causing conversion to and from C float to lose precision. It is best practice to test for small value ranges around a value rather than rely on exact values.
  • AVRational values map to and from a Javascript array containing two integer numbers. For example, C value (AVRational){1,25} maps to Javascript value [1,25].
  • C char * and const char * map to and from Javascript's string type.
  • FFmpeg enumerations are converted to Javascript strings representing their name. FFmpeg provides its own utility functions for converting many of its enumerations to and from strings, for example av_color_space_name and av_color_space_from_name.
  • FFmpeg flags are converted to and from a Javascript object of Boolean-valued properties, where each property is named to match the distinguishing part of the FFmpeg #define macro name. For example, the corrupt and discard flags of an AVFrame flags property become { CORRUPT: false, DISCARD: false } in Javascript. When setting, use an object, e.g. to update the CORRUPT flag, do frame.flags = { CORRUPT: true };.
  • AVBufferRef and other blobs of binary data are represented by Node.js Buffers. Wherever possible, the data is not copied. For example, when reading an AVBuffer from Javascript, a new reference to the underlying AVBuffer is created and a view of the raw data made available via an external buffer. The reference is held until the Javascript buffer is garbage collected. Conversely, when a Javascript buffer is to be read from an FFmpeg structure, the AVBufferRef structure holds a V8 reference to the underlying data of the Javascript buffer. The Javascript buffer is kept alive until both of any AV processing functions and Javascript are finished and the buffers are unreferenced.
  • Binary data blobs of type uint8_t *, such as extradata or AVPacketSideData.data are assumed to be small and easy to copy. Javascript getters make a copy of the underlying data and return a Buffer. The setters create a copy of the data in a buffer and use it to set the underlying value. Therefore, to modify the data, it must be read via the getter, modified and written back via the setter.
  • AVDictionary metadata and private data values have a natural mapping to Javascript objects as keys to property names and their value pair. Always set dictionary-based and private data values using an object. Dictionary values are replaced in their entirety. For private data, only the properties contained in the update object will be modified.

Demuxing

The process of demuxing (de-multiplexing) extracts time-labelled packets of data contained in a media stream or file. FFmpeg provides a diverse range of demuxing capability with support for a wide range of input formats and protocols (beamcoder.protocols()).

To see a list and details of all the available demuxer input formats:

let dms = beamcoder.demuxers();

The output is an object where each property key is the name of a demuxer and each value is an object describing the input format. For example, to find the demuxer for .mp4 files:

Object.values(dms).filter(x => x.extensions.indexOf('mp4') >= 0);
[ { type: 'InputFormat',
    name: 'mov,mp4,m4a,3gp,3g2,mj2',
    long_name: 'QuickTime / MOV',
    mime_type: '', // Generally not available for demuxers - see equivalent muxers
    extensions: 'mov,mp4,m4a,3gp,3g2,mj2',
    flags:
     { NOFILE: true, // Demuxer will manage IO operations
       NEEDNUMBER: false, // Needs '%d' in filename
       SHOW_IDS: true, // Show format stream IDs numbers.
       GENERIC_INDEX: true, // Use generic index building code
       TS_DISCONT: true, // Format allows timestamp discontinuities.
       NOBINSEARCH: true, // Format does not allow to fall back on binary search via read_timestamp
       NOGENSEARCH: true, // Format does not allow to fall back on generic search
       NO_BYTE_SEEK: false, // Format does not allow seeking by byte
       SEEK_TO_PTS: false }, // Seeking is based on PTS
    raw_codec_id: 0,
    priv_data_size: 400,
    priv_class: null } ]

The easiest way to create a demuxer is with a filename or URL, for example to open a transport stream containing and HEVC encoding of the Big Buck Bunny short movie in file bbb_1080p_c.ts in a movie sub-directory:

let tsDemuxer = await beamcoder.demuxer('file:media/bbb_1080p_c.ts');

The demuxer operation performs file system and/or network access and so is asynchronous. On successful resolution after the contents of the file or stream has been probed, the value is a Javascript object describing the contents of the media input. Here is a summary of the created demuxer:

{ type: 'demuxer',
  iformat:
   { type: 'InputFormat',
     name: 'mpegts',
     \* ... *\ },
  ctx_flags: { NOHEADER: false, UNSEEKABLE: false },
  streams:
  [ { type: 'Stream',
      index: 0,
      id: 301,
      time_base: [ 1, 90000 ], /* ... */
      codecpar:
       { type: 'CodecParameters',
         codec_type: 'video',
         codec_id: 173,
         name: 'hevc',
         codec_tag: 'HEVC',
         format: 'yuv420p', /* ... */
         width: 1920,
         height: 1080,
        /* ... */ } },
    { type: 'Stream',
      index: 1,
      id: 302,
      time_base: [ 1, 90000 ], /* ... */
      codecpar:
       { type: 'CodecParameters',
         codec_type: 'audio',
         codec_id: 86018,
         name: 'aac', /* ... */
         format: 'fltp', /* ... */
         channel_layout: 'stereo',
         channels: 2,
         /* ... */ } } ],
  url: '../media/bbb_1080p_c.ts',
  start_time: 80000,
  duration: 596291667,
  bit_rate: 2176799, /* ... */
  stream: [Function],
  read: [Function: readFrame],
  seek: [Function: seekFrame] }

From this object dump, it is possible to determine that the file contains two streams:

  1. an HD video stream encoded with H.265/HEVC;
  2. a stereo audio stream encoded with AAC.

The duration property is measured in microseconds (AV_TIME_BASE), so is approximately 596 seconds or 9m56s.

For formats that require additional metadata, such as the rawvideo format, it may be necessary to pass additional information such as image size or pixel format. To do this, pass in an options object with a url property for the filename(s) (may contain %d for a sequence of numbered files) and options property for the values. For example:

let rawDemuxer = await beamcoder.demuxer({
  url: 'file:movie/bbb/raw_pictures_%d.rgb',
  options: {
    video_size: '640x480',
    pixel_format: 'rgb24'
  }
});

Reading data packets

To read data from the demuxer, use the read method of a demuxer-type object, a method that takes no arguments. This reads the next blob of data from the file or stream at the current position, where that data could be from any of the streams. Typically, a packet is one frame of video data or a data blob representing a codec-dependent number of audio samples. Use the stream_index property of returned packet to find out which stream it is associated with and dimensions including height, width or audio sample rate. For example:

let wavDemuxer = await beamcoder.demuxer('file:my_audio.wav');
let packet = {};
while (packet != null) {
  packet = await wavDemuxer.read();
  if (packet &&
      wavDemuxer.streams[packet.stream_index].codecpar.codec_type === 'video') {
    // Do something with the image data
  }
}

The read method is asynchronous and returns a promise. The promise resolves to an object of type Packet if it succeeds with data, or a null value at the end of the file. If an error occurs, the promise rejects with an error message. An example of a successful read from a WAVE file is shown below:

{ type: 'Packet',
  pts: 2792448, // presentation timestamp, measured in stream timebase
  dts: 2792448, // decode timestamp, measured in stream timebase
  data: // the raw data of the packet
   <Buffer fd ff fe ff 02 00 01 00 ff ff 00 00 00 00 ff ff 00 00 02 00 02 00 fe ff fd ff 01 00 04 00 00 00 fb ff ff ff 04 00 01 00 fe ff ff ff 01 00 01 00 00 00 ... >,
  size: 4160, // the size of the raw data
  stream_index: 0, // the stream index of the stream this packet belongs to
  flags:
   { KEY: true, // Packet represents a key frame
     CORRUPT: false, // Corruption detected
     DISCARD: false, // Can be dropped after decoding
     TRUSTED: false, // Packet from a trusted source
     DISPOSABLE: false }, // Frames that can be discarded by the decoder
  duration: 1024, // Wrt the stream timebase
  pos: 11169836 } // Byte offset into the file

Seeking

Beam coder offers FFmpeg's many options for seeking a particular frame in a file, either by time reference, frame count or file position. To do this, use the seek method of a demuxer-type object with an options object to configure the operation.

To seek forward to a keyframe in a given stream or file at a given timestamp:

await demuxer.seek({ stream_index: 0, timestamp: 2792448 });

The timestamp is the presentation timestamp of the packet measured in the timebase of the stream, which is [1, 48000] in this example. To seek based on elapsed time from the beginning of the primary stream (as determined by FFmpeg, normally the first video stream where available), use the time property:

await demuxer.seek({ time: 58.176 });

Another form of seek is to use a byte offset position into the file:

await demuxer.seek({ pos: 11169836 });

The final form of seeking supported is by number of frames into a given stream:

await demuxer.seek({ frame: 42, stream_index: 0});

All successful seek calls resolve to a null value, or reject if the kind of seek is not supported or an error occurs. To access the seeked-for data, call read to get the next frame. Note that if you seek beyond the end of the file or stream, the seek call resolves OK and the next read operation resolves to null.

The seek operation has two additional flags that can be specified. The backward Boolean-valued property is interpreted as find the nearest key frame before the timestamp when set and find the nearest keyframe after the timestamp when not. By default, it is set to true. The any Boolean-valued property enables seeking to both key and non-key frames. For example:

await demuxer.seek({ frame: 31, stream_index: 0, backward: false, any: true});

Demuxer stream

Beam coder offers a Node.js Writable stream interface to a demuxer, allowing source data to be streamed to the demuxer from a file or other stream source such as a network connection.

To create a Writable stream interface with for example a 64kbyte threshold, use:

let demuxerStream = beamcoder.demuxerStream({ highwaterMark: 65536 });

This stream can then have data written to it or piped from a source stream

fs.createReadStream('file:media/bbb_1080p_c.ts').pipe(demuxerStream);

Once the source data is connected to the stream, the demuxer can be created:

let demuxer = await demuxerStream.demuxer();

This function will return a promise that will resolve when it has determined sufficient format details by consuming data from the source. The promise will wait indefinitely until sufficient source data has been provided.

Decoding

Decoding is the process of taking a stream of compressed data in the form of packets and converting it into uncompressed frames. In general, to decode an interleaved (multiplexed) media file, you need a decoder for each of the video and the audio streams. For the purpose of keeping the examples simple in this section, only a single stream is decoded. However, it is possible to set up more than one decoder - say for related video and audio streams - and run them asynchronously and in parallel, i.e. to decode the video and audio required to present a frame.

To see a list of available decoders, use:

let decs = beamcoder.decoders();

As with the demuxers, the result is an object where the keys are the names of the decoders and the values are objects describing the codec. This includes the codec type (video, audio, subtitle), a descriptor for the family of codecs, some capability flags, supported profiles and more. Here are some examples of querying the available decoders:

// Find a decoder that deals with H.264 / AVC
decs['h264'];
{ type: 'Codec',
  name: 'h264',
  long_name: 'H.264 / AVC / MPEG-4 AVC / MPEG-4 part 10',
  codec_type: 'video',
  id: 27, /* ... */ }

// Find all other decoders that can decode H.264
Object.values(decs).filter(x => x.id === 27).map(x => x.name);
[ 'h264', 'h264_qsv', 'h264_cuvid' ]
// Note: h264_qsv and h264_cuvid are wrappers to hardware-accelerated decoders
//       Appearance in the list does not imply that the codec is supported by current hardware.

// Find all decoders claiming support for the MP3 format
Object.values(decs)
  .filter(x => x.long_name.indexOf('MP3') >= 0)
  .map(x => ({ name: x.name, long_name: x.long_name }));
[ { name: 'mp3float', long_name: 'MP3 (MPEG audio layer 3)' },
  { name: 'mp3', long_name: 'MP3 (MPEG audio layer 3)' },
  { name: 'mp3adufloat',
    long_name: 'ADU (Application Data Unit) MP3 (MPEG audio layer 3)' },
  { name: 'mp3adu',
    long_name: 'ADU (Application Data Unit) MP3 (MPEG audio layer 3)' },
  { name: 'mp3on4float', long_name: 'MP3onMP4' },
  { name: 'mp3on4', long_name: 'MP3onMP4' } ]

// List all audio decoders
Object.values(decs).filter(x => x.codec_type === 'audio').map(x => x.name);
[ 'comfortnoise', 'dvaudio', '8svx_exp', '8svx_fib', 's302m', 'sdx2_dpcm',
  'aac', 'aac_fixed', 'aac_latm' /* ... */ ]

See the FFmpeg Codecs Documentation for more details.

Decoder

To create an instance of a decoder, request a decoder from beam coder, specifying either the decoder's name, a codec_id, or by providing a demuxer and a stream_id. For example:

// Create a decoder by name - note this is synchronous
let decoder = beamcoder.decoder({ name: 'h264' });
// or for the first choice codec in the H.264 family
let decoder = beamcoder.decoder({ codec_id: 27 });
// Alternatively, use a demuxer and a stream_index
//   The codec parameters of the streams will be used to set the decoding parameters
let tsDemux = await beamcoder.demuxer('media/bbb_1080p_c.ts');
let decoder = beamcoder.decoder({ demuxer: tsDemux, stream_index: 0 });

Other properties of the decoder can be provided on initialisation and may be required in certain cases. For example, the width and height of the video. These will override the default values. Note that certain values for the decoder may not be available until a the first few packets have been decoded.

let decoder = beamcoder.decoder({ name: 'h264', width: 1920, height: 1080 });

A decoder has many properties. These can be set before decoding in the usual way for a Javascript object. Some of the properties are more appropriate for encoding but are present for information. Some properties can only be set by libav*, others can only be set by the user, some both. Follow the AVCodecContext FFmpeg documentation for details.

Decode

To decode an encoded data packet and create an uncompressed frame (may be a frames-worth of audio), use the asynchronous decode method of a decoder. Decoders may need more than one packet to produce a frame and may subsequently produce more than one frame per packet. This is particularly the case for long-GOP formats.

let packet = demuxer.read();
while (packet != null) {
  let dec_result = await decoder.decode(packet);
  // dec_result.frames - possibly empty array of frames for further processing
  // dec_result.total_time - microseconds that the operation took to complete
  // Get the next packet, e.g. packet = demuxer.read();
}

As long as decoding was successful, the decode operation resolves to an object containing the total_time (measured in microseconds) that the operation took to execute, and an array of decoded frames that are now available. If the array is empty, the decoder has buffered the packet as part of the process of producing future frames. Frames are delivered in presentation order.

It is possible to pass more than one packet at a time to the decoder, either as an array of packets or a list of arguments:

// Array of packets
let dec_result = await decoder.decode([packet1, packet2, packet3 /* ... */ ]);
// List of packets as arguments
let dec_result = await decoder.decode(packet1, packet2, packet3 /* ... */ );

Creating packets

Packets for decoding can be created without reading them from a demuxer. For example:

beamcoder.packet({ pts: 43210, dts: 43210, data: Buffer.from(...) });

Packet data buffers are shared between C and Javascript so can be written to and modified without having to write the buffer back into the packet.

Flush decoder

Once all packets have been passed to the decoder, it is necessary to call its asynchronous flush() method. If any frames are yet to be delivered by the decoder, they will be provided in the resolved value.

let flush_result = await decoder.flush();
// flush_result.frames - array of any remaining frames to be decoded
// flush_result.total_time - microseconds taken to execute the flush operation

Call the flush operation once and do not use the decoder for further decoding once it has been flushed. The resources held by the decoder will be cleaned up as part of the Javascript garbage collection process, so make sure that the reference to the decoder goes out of scope.

Filtering

Filtering is the process of taking streams of uncompressed data in the form of frames and processing them through a chain of connected filters in order to produce modified uncompressed data again in the form of frames. Filtering takes place on a single type of stream, either audio or video. Filtering chains may have multiple inputs and/or multiple outputs.

To see a list of available filters, use:

let filters = beamcoder.filters();
{ abench:
   { name: 'abench',
     description: 'Benchmark part of a filtergraph.',
     inputs: [ [Object] ],
     outputs: [ [Object] ],
     priv_class: { type: 'Class', class_name: 'abench', options: [Object] },
     flags:
      { DYNAMIC_INPUTS: false,
        DYNAMIC_OUTPUTS: false,
        SLICE_THREADS: false,
        SUPPORT_TIMELINE_GENERIC: false,
        SUPPORT_TIMELINE_INTERNAL: false } },
  acompressor:
   { name: 'acompressor',
     description: 'Audio compressor.',
     inputs: [ [Object] ],
     outputs: [ [Object] ],
     priv_class:
      { type: 'Class', class_name: 'acompressor', options: [Object] },
     flags:
      { DYNAMIC_INPUTS: false,
        DYNAMIC_OUTPUTS: false,
        SLICE_THREADS: false,
        SUPPORT_TIMELINE_GENERIC: false,
        SUPPORT_TIMELINE_INTERNAL: false } },
  /* ... */
}

Filterer

To create an instance of a filterer, request a filterer from beam coder, specifying the stream type, input parameters, output parameters and a filter definition string. The input parameters will typically be read from the selected demuxer stream object.

let a_filterer = await beamcoder.filterer({
  filterType: 'audio',
  inputParams: [
    {
      sampleRate: audStream.codecpar.sample_rate,
      sampleFormat: audStream.codecpar.format,
      channelLayout: 'mono',
      timeBase: audStream.time_base
    }
  ],
  outputParams: [
    {
      sampleRate: 8000,
      sampleFormat: 's16',
      channelLayout: 'mono'
    }
  ],
  filterSpec: 'aresample=8000, aformat=sample_fmts=s16:channel_layouts=mono'
});

let v_filterer = await beamcoder.filterer({
  filterType: 'video',
  inputParams: [
    {
      width: vidStream.codecpar.width,
      height: vidStream.codecpar.height,
      pixelFormat: vidStream.codecpar.format,
      timeBase: vidStream.time_base,
      pixelAspect: vidStream.sample_aspect_ratio,
    }
  ],
  outputParams: [
    {
      pixelFormat: 'yuv422'
    }
  ],
  filterSpec: 'scale=1280:720'
});

The filter format string describes the filters to use, their parameters and the connections between them. See the filtering documentation from FFmpeg.

In order to support multiple inputs and outputs, name parameters are used. As above if there is only a single input or a single output, the name is optional. The name parameters when present must match the names in the filter string:

let v_filterer = await beamcoder.filterer({
  filterType: 'video',
  inputParams: [
    {
      name: 'in0:v',
      width: vidStream0.codecpar.width,
      height: vidStream0.codecpar.height,
      pixelFormat: vidStream0.codecpar.format,
      timeBase: vidStream0.time_base,
      pixelAspect: vidStream0.sample_aspect_ratio,
    },
    {
      name: 'in1:v',
      width: vidStream1.codecpar.width,
      height: vidStream1.codecpar.height,
      pixelFormat: vidStream1.codecpar.format,
      timeBase: vidStream1.time_base,
      pixelAspect: vidStream1.sample_aspect_ratio,
    }
  ],
  outputParams: [
    {
      name: 'out0:v',
      pixelFormat: 'yuv422'
    }
  ],
  filterSpec: '[in0:v] scale=1280:720 [left]; [in1:v] scale=640:360 [right]; [left][right] overlay=format=auto:x=640 [out0:v]'
});

If the specified output parameters differ from the format produced by the last filter specified in the filter string then a resample or scaler filter will be automatically inserted.

The properties of the resolved filterer object can be examined through the object's graph property.

{ type: 'filterGraph',
  filters:
   [ { type: 'FilterContext',
       filter: [Object],
       name: '0:a',
       /* ... */
       priv: [Object],
       /* ... */ },
     { type: 'FilterContext',
       filter: [Object],
       name: 'out',
       /* ... */
       priv: [Object],
       /* ... */ },
     { type: 'FilterContext',
       filter: [Object],
       name: 'Parsed_aresample_0',
       /* ... */
       priv: [Object],
       /* ... */ },
     { type: 'FilterContext',
       filter: [Object],
       name: 'Parsed_aformat_1',
       /* ... */
       priv: [Object],
       /* ... */ } ],
  /* ... */
  dump: [Function] }

An ascii string represention of the filter graph can be accessed via:

let dump = filterer.graph.dump();

The parameters of an individual filter are available via that filter's priv property. These parameters can be read from Javascript and some of them are able to be set (a per-filter implementation decision) by providing a complete value. For example to set the scale filter width parameter:

let scaleFilter = filterer.graph.filters.find(f => 'scale' === f.filter.name); // find the first 'scale' filter
scaleFilter.priv = { width: 1000 }; // only the included properties are changed, other existing properties are unaltered
// scaleFilter.priv.width = 1000; - this will not work !!

Filter

To filter uncompressed frames and create uncompressed result frames (which may each be frames-worth of audio), use the filter method of a filterer, passing arrays of objects, one per filter input, each with a name string property and a frames property that contains an array of frame objects, for example from the output of a decoder:

let filtFrames = await filterer.filter([
  { name: 'in0:v', frames: frames0 },
  { name: 'in1:v', frames: frames1 },
]);

The name parameters must match the input parameters in the filterSpec string used in the filterer setup.

If the filter has only one input then a simplification is available that takes the array of frame objects directly and will apply a default name for the filter input. This name will match a filter specification that doesn't name its inputs:

let filtFrames = await filterer.filter(frames);

The result is an array containing objects with a frames property which is an array of frame objects.

Multiple output objects appear in the output array with each object having a name property that matches the output name property defined in the filterSpec string. For single input and single output filters the name property is optional, though if the filter string has defined a name the same name must be used.

The output array objects also contain a total_time property which logs the time the operation took to complete.

Filters do not need to be flushed.

Encoding

Encoding is the process of taking a stream of uncompressed data in the form of frames and converting them into coded packets. Encoding takes place on a single type of stream, for example audio or video. Downstream, encoded date maybe combined into outputs by a muxer.

To see a list of available encoders, use:

let encs = beamcoder.encoders();

The encoders are listed in with the same form and structure as the decoders and can be queried in the same way.

// Find all encoders that can encode H.264
Object.values(encs).filter(x => x.id === 27).map(x => x.name);
[ 'libx264', 'libx264rgb', 'h264_amf', 'h264_nvenc', 'h264_qsv', 'nvenc', 'nvenc_h264' ]
etc..

Encoder

To create an instance of an encoder, request an encoder from beam coder, specifying either the codec's name, or a codec_id.

// Create an encoder by name of name of a codec family
let v_encoder = beamcoder.encoder({ name: 'h264' });
// In this case, the 'libx264' software codec will be selected.

// Create an encoder by codec identifier - in this case for 'aac' audio
let a_encoder = beamcoder.encoder({ codec_id: 86018 });

Additional properties may be set on the encoder object either during construction in the options object or using setters. See the AVFormatContext documentation from FFmpeg for details of the parameters. In general, the user sets up all the configuration parameters prior to starting the encode process. Note hat libav* may set some of the parameters during encoding.

Each codec may have additional private data settings and these are available in the encoder's priv_data property. These can be read and set from Javascript but note that the value can only be set by providing a complete value. For example, to set the libx264 encoder to the slow preset:

v_encoder.priv_data = { preset: 'slow' }; // Only the included properties change

The following will not work:

v_encoder.priv_data.preset = 'slow';

Encode

To encode an uncompressed frame and create a compressed packet (may be a frames-worth of audio), use the encode method of an encoder. Encoders may need more than one frame to produce a packet and may subsequently produce more than one packet per frame. This is particularly true for long-GOP formats.

// Get or make the first frame of data, store in frame
while (frame != nullptr) {
  let enc_result = await encoder.encode(frame);
  // enc_result.packets - possibly empty array of packets for further processing
  // enc_result.total_time - microseconds that the operation took to complete
  // Get or make the next frame of data
}

As with decoding, it is possible to encode more than one frame at a time. Pass in an array of frames or a sequence of frames as arguments to the method:

enc_result = await encoder.encode([ frame_1, frame_2, ...  ]);
enc_result = await encoder.encoder(frame_1, frame_2, ... );

Creating frames

Frames for encoding or filtering can be created from Javascript as follows:

beamcoder.frame({ pts: 43210, width: 1920, height: 1920, format: 'yuv420p',
  data: [ Buffer.from(...plane_1...), Buffer.from(...plane_2...), ... ] });

In general, frame data is planar, e.g. split into separate buffers for Red (R), Green (G) and Blue (B) components or Chrominance (Y) and Luminance (UV) components. Alpha (A) channel, if present, is always the last component. For details of the planes expected for each pixel format, call:

beamcoder.pix_fmts();

Planar audio representations - those with a p in their name - use planes to represent data for each audio channel. For details of sample formats, call:

beamcoder.sample_fmts();

Beam coder exposes some of FFmpeg's ability to calculate the size of data buffers. If you pass width, height and format properties for video frames, or channels/channel_layout, sample_rate and format for audio frames, as options to the frame constructor then the linesize array (number of bytes per line per plane) is computed. For video, multiply each value by the height to get the minimum buffer size for the plane. For audio, the first element of the array is the buffer size for each plane.

To use the linesize numbers to automatically allocate buffers of the correct size, call alloc() after the factory method. For example:

let f = beamcoder.frame({ width: 1920, height: 1080, format: 'yuv422p' }).alloc();

Note that when creating buffers from Javascript, FFmpeg recommends that a small amount of headroom is added to the minimum length of each buffer. The minimum amount of padding is exposed to Javascript as constant:

beamcoder.AV_INPUT_BUFFER_PADDING_SIZE

Flush encoder

Once all frames have been passed to the encoder, it is necessary to call the asynchronous flush() method. If any packets are yet to be fully encoded or delivered by the encoder, they will be completed and provided in the resolved value.

let flush_result = await encoder.flush();
// flush_result.frames - array of any remaining packets to be encoded
// flush_result.total_time - microseconds taken to execute the flushing process

Call the flush method once and do not use the encoder for further encoding once it has been flushed. The resources held by the encoder will be cleaned up as part of the Javascript garbage collection process, so make sure that the reference to the encoder goes out of scope.

Muxing

Muxing (multiplexing) is the operation of interleaving media data from multiple streams into a single file or stream, the opposite process to demuxing. In its simplest form, a single stream is written to a file, adding any necessary headers, padding or trailing data according to the file format. For example, writing a WAVE file involves writing a header followed by the PCM audio data.

For more complex formats, such as an MP4 file, video and audio data packets with similar timestamps are written to the same section of the output file. When the file is subsequently read, video and audio that needs to be presented together can be read from the same part of the file.

The muxing procedure has five basic stages:

  1. Choosing the output format, creating the muxer and setting up the output stream metadata.
  2. Opening the file or streams for writing.
  3. Writing the header metadata.
  4. In a loop, writing packets and/or frames.
  5. Writing the trailer and closing the file or stream.

Here is an example of reading and writing a WAVE file as an illustration:

let demuxer = await beamcoder.demuxer('file:input.wav');
let muxer = beamcoder.muxer({ filename: 'file:test.wav' });
let stream = muxer.newStream(demuxer.streams[0]);
await muxer.openIO();
await muxer.writeHeader();
let packet = {};
for ( let x = 0 ; x < 100 && packet !== null ; x++ ) {
  packet = await demuxer.read();
  await muxer.writeFrame(packet);
}
await muxer.writeTrailer();

Muxer

A first stage of creating a muxer is to select an output format. The complete list of output formats can be queried as follows:

let mxs = beamcoder.muxers();

Each muxer has an extensions property that is a comma-separated list commonly used file extensions for the output format. This can be searched (see demuxing example), or alternatively beam coder exposes FFmpeg's capability to guess an output format from a filename, format name or mime type. For example, for output format of:

of = beamcoder.guessFormat('mpegts'); // Finds the output format named 'mpegts'
of = beamcoder.guessFormat('image/png'); // Output format for PNG image data
of = beamcoder.guessFormat('fred.jpeg'); // Output format for '.jpeg' file extension

To create a muxer, user an options object and do one of the following:

  • set a format_name or name property to that of an output format;

  • set a filename property to a filename with extension to be used to guess the output format;

  • set an oformat property to an output format object retrieved with, say, beamcoder.muxers() or beamcoder.guessFormat().

    muxer = beamcoder.muxer({ format_name: 'mp4' }); muxer = beamcoder.muxer({ filename: 'example.wav' }); muxer = beamcoder.muxer({ oformat: of });

For details of the properties of a muxer, see the FFmpeg AVFormatContext documentation.

The next step is to add the streams for the format. This can only be done using the newStream() method of a muxer and must be done in order. The steps are:

  1. Create the stream with a codec name.
  2. Set the time_base for the stream.
  3. Update the codec parameters for the stream, ensuring at least width, height and format (pixel format) are set for video streams and sample_rate, channels and/or channel_layout and format (sample format) are set of audio streams.

This procedure is illustrated in the following code:

let muxer = beamcoder.muxer({ filename: 'test.wav' }); // Filename set muxer's 'url' property
let stream = muxer.newStream({
  name: 'pcm_s16le',
  time_base: [1, 48000 ],
  interleaved: false }); // Set to false for manual interleaving, true for automatic
Object.assign(stream.codecpar, { // Object.assign copies over all properties
  channels: 2,
  sample_rate: 48000,
  format: 's16',
  channel_layout: 'stereo',
  block_align: 4, // Should be set for WAV
  bits_per_coded_sample: 16,
  bit_rate: 48000*4
});

Note that newStream() adds the stream to the muxer automatically using the next available stream index. A reference to the created stream is returned by newStream() for convenience. For details of the properties of stream, see FFmpeg's AVStream and AVCodecParameters documentation.

An alternative way to create muxers, e.g. for a transmuxing process (direct connection from demuxer to muxer to rewrap content without transcoding), a new muxer stream can be created with a demuxer. An output format will be found to match the input format and the streams and their codec parameters will be copied over. For example:

let muxAudioStream = muxer.newStream(demuxer.streams[1]);

Opening the output

The next step is to open the output file or stream. If a filename or URL has been provided with the creation of the muxer using the filename property, the output can be opened with the asynchronous openIO() method of the muxer.

Alternatively, pass in an options object containing the filename or url. The options object can also itself contain an options property, an object, to further configure the protocol with private data. See the FFmpeg protocol documetation for information about protocol-specific options. In the simplest case, where a filename was provided when the muxer was created, simply call the method without any arguments:

await muxer.openIO();

An example of opening the muxer's output for a WAVE file with options:

await muxer.openIO({
  url: 'file:test2.wav', // Update URL
  options: { // Protocol private options
    blocksize: 8192
  }
});

On success, the returned promise resolves to undefined or, if some of the options could not be set, an object containing an unset property detailing which of the properties could not be set.

Note: An outstanding task is to provide some kind of getter for the protocol private data through the Javascript API.

Writing the header

The next stage is to write the header to the file. This must be done even for formats that don't have a header as part of the internal structure as this step also initializes the internal data structures for writing. This is the simplest example of calling the asynchronous writeHeader() method of the muxer:

await muxer.writeHeader();

As with openIO(), it is possible to pass in private data options at this stage that set the private data fields of the muxer:

await muxer.writeHeader({
  options: { // Private class options
    write_peak: 'on', // Write Peak Envelope chunk - enum of 'off', 'on', 'only'.
    peak_format: 2, // The format of the peak envelope data (1: uint8, 2: uint16).
    write_bext: false // Write BEXT chunk.
  }
});

Subsequently to writing the header, the options can be viewed and modified through the priv_data property of the muxer. For details of the available options per format, see the FFmpeg muser documentation or the associated OutputFormat object which is the oformat property of the muxer. Specifically:

console.log(muxer.oformat.priv_class.options);

In some cases, it is necessary to initialize the structures of the muxer before writing the header. In this case, call the asynchronous initOutput() method of the muxer first. This method can also take options to initialise muxer-specific parameters. Further configure the initialized muxer and then call writeHeader() before writing any packets or frames.

On success, the writeHeader() and initOutput() methods both resolve to an object that says where the initialisation of the muxer took place, with details of any of the options that were unset. For example:

{ INIT_IN: 'WRITE_HEADER', unset: { write_bext: false } }

Writing packets and frames

To write the actual media data to the file, it is necessary to send frames and packets containing the data of the media streams. Packets must contain their stream index, frames must be sent with alongside their stream index and both must have timestamps measured in the time_base of their stream. (With the exception of streams that don't have a time_base, in which case set beamcoder.AV_NOPTS_VALUE.)

The muxer has an interleaved property that determines whether beam coder asks FFmpeg to do the interleaving of streams or whether the user is taking charge. Set the property to true (default) for FFmpeg to do the work, otherwise set it to false. Set this property by this point and don't change it. From here on in, the approaches cannot be mixed and matched.

Send packets or frames from across all the streams with similar presentation timings in the stream in a round robin-like loop, sending more packets per iteration for streams where the packets/frames have shorter durations. In interleaving mode, FFmpeg builds up a buffer of time-related packets and writes them when it has sufficient information to cover a given time segment.

Packets or frames are sent to the muxer with its asynchronous writeFrame() method. The method expects a single argument that is either a packet or an options object. If an options object, provide either a single packet property, or a frame and a stream_index property. Here are some examples:

await muxer.writeFrame(packet);
await muxer.writeFrame({ packet: packet });
await muxer.writeFrame({ frame: frame, stream_index: 0 });

The writeFrame() promise resolves to undefined on success, otherwise the promise rejects with an error.

Writing the trailer

The trailer is the end of the file or stream and is written after the muxer has drained its buffers of all remaining packets and frames. Writing the trailer also closes the file or stream. Use the asynchronous writeTrailer() method. It takes no arguments:

await muxer.writeTrailer();

On success, the promise resolves to an undefined value. Do not try to write other data to the muxer after calling this method. Any other resources held by the muxer will be released by Javascript garbage collection.

To abandon the muxing process and forcibly close a file or stream without completing it, call the synchronous forceClose() method of the muxer. This assumes that any result of the muxing process is to be left in an incomplete and invalid state.

Muxer stream

Beam coder offers a Node.js Readable stream interface to a muxer, allowing muxed data to be streamed out to a file or other stream destination such as a network connection.

To create a Readable stream interface with for example a 64kbyte threshold, use:

let muxerStream = beamcoder.muxerStream({ highwaterMark: 65536 });

This stream can then have data read or piped from it:

muxerStream.pipe(fs.createWriteStream('test.wav'));

Once the stream is initialised, the muxer can be created and used as above:

let muxer = muxerStream.muxer({ format_name: 'wav' });

The muxer async methods such as writeFrame return a promise that will resolve when the Readable stream has bufferred the packet. If the Readable stream is not flowing or the buffer is full the promise will wait indefinitely.

Codec parameters

Another mechanism of passing parameters from demuxers to decoders to encoders and then muxers is to use codec parameters. These are a set of parameters that are codec-type specific, e.g. common to all codec implementations of a codec family such as H.264. They can be used to uniquely identify and represent the kind of codec of a stream and its dimensions and properties.

Here are the various places that codec parameters can be found:

  • In previous sections, examples of codec parameters as the codecpar property of a demuxer's streams were provided. Codec parameters can be referenced from streams and used to set up decoders.
  • In previous examples, codec parameters were used to set the properties of a muxer's new streams.
  • Codec parameters have a factory method beamcoder.codecParameters(). Pass in a name parameter and other property values in an options object to configure the parameters on construction, or set these properties using setter methods.
  • Instead of using a codec name, decoders and encoders can be created from codec parameters. Use the params property of the factory method.
  • A copy of codec parameters can be extracted from a decoder or a decoder using their extractParameters() method.
  • Codec parameters can be used to override the currently set properties for a decoder or an encoder using the useParams() method. Note that the codec_type and codec_id must be the same as that already set for decoder or encoder.

Here are some contrived examples of using codec parameters:

let demuxer = await beamcoder.demuxer('file:bbb_1080p_c.ts');
let videoParams = demuxer.streams.find(x => x.codecpar.codec_type === 'video').codecpar;
// Set up a decoder using demuxer stream parameters
let decoder = beamcoder.decoder({ params: videoParams });

let encParams = decoder.extractParams(); // Make a copy of the decoder paramerers
encParams.name = 'h264'; // Change the codec but keep the other parameters ...
let encoder = beamcoder.encoder({ params: videoParams }); // ... to set up an encoder

let muxer = beamcoder.muxer({ filename: 'file:bbb.mp4' });
let stream = muxer.newStream({ // ... to set up a new stream in the muxer
  name: encParams.name,
  time_base: [1, 90000],
  codecpar: encParams
});

Care must be taken not to change codec parameters by side effects, for example using the same codec parameters in a transcoder where the decoder parameters are used to set up the encoder without copying them (extracting them) first. Modifications to the shared codec parameters object could effect both the decoder and encoder.

Reactive streams

In order to facilitate more complex decoding, filtering, encoding and muxing operations an experimental Reactive Streams implementation of a configurable pipeline of FFmpeg components is available.

The pipeline supports multiple sources, each with a demuxer and decoder, feeding into a filter that will process them to one or more outputs that are each encoded and sent to a muxer stream. Each FFmpeg processing element works asynchronously while the pipeline keeps all the input, processing and output streams in step as required according to their timestamps. Small buffers are maintained between each processing element and back pressure is used to ensure that resources are kept reasonably constrained.

A parameters object is set up to describe the sources, filters and mux streams that are desired and passed to a function that will create the necessary components, connect them together and start streaming. Each sources object maps to filter inputs in array order and the filter specification must use input names in the form in0:v where the number increments according to the sources array ordinal and the 'v' refers to video and is replaced by 'a' for audio filters. Similarly each streams object expects to receive an output from the filter with a name in the form out0:v.

Below is an example parameters object describing a simple transformation:

const urls = [ 'file:../../Media/big_buck_bunny_1080p_h264.mov' ];
const spec = { start: 0, end: 48 };

const params = {
  video: [
    {
      sources: [
        { url: urls[0], ms: spec, streamIndex: 0 }
      ],
      filterSpec: '[in0:v] scale=1280:720, colorspace=all=bt709 [out0:v]',
      streams: [
        { name: 'h264', time_base: [1, 90000], encoderName: 'libx264',
          codecpar: {
            width: 1280, height: 720, format: 'yuv422p', color_space: 'bt709',
            sample_aspect_ratio: [1, 1]
          }
        }
      ]
    }
  ],
  audio: [
    {
      sources: [
        { url: urls[0], ms: spec, streamIndex: 2 }
      ],
      filterSpec: '[in0:a] aformat=sample_fmts=fltp:channel_layouts=mono [out0:a]',
      streams: [
        { name: 'aac', time_base: [1, 90000], encoderName: 'aac',
          codecpar: {
            sample_rate: 48000, format: 'fltp', frame_size: 1024,
            channels: 1, channel_layout: 'mono'
          }
        }
      ]
    },
  ],
  out: {
    formatName: 'mp4',
    url: 'file:temp.mp4'
  }
};

The parameters are split into video and audio arrays each with sources, a filter specification string and output streams. A optional time specification object is also provided so that a section of the source can be extracted.

To start processing, simply pass the params object to the source creator, which will make the demuxers and source streams, then to the stream creator, which makes the output:

await makeSources(params);
await makeStreams(params);

The processing will begin immediately and will continue until the time specification has been completed or the end is reached of any of the sources.

The UV_THREADPOOL_SIZE environment variable (see above) can be used to control the number of asynchronous processes that are progressing at once and hence the impact on the CPU load in the computer that is running the program.

Status, support and further development

Although the architecture of the aerostat beam coder is such that it could be used at scale in production environments, development is not yet complete. In its current state, it is recommended that this software is used in development environments, primarily for building prototypes. Future development will make this more appropriate for production use.

The developers of beam coder aimed to find a balance between being a faithful mapping of FFmpeg to Javascript while creating a Javascript API that is useful and easy to use. This may mean that certain features of FFmpeg are