npm package discovery and stats viewer.

Discover Tips

  • General search

    [free text search, go nuts!]

  • Package details

    pkg:[package-name]

  • User packages

    @[username]

Sponsor

Optimize Toolset

I’ve always been into building performant and accessible sites, but lately I’ve been taking it extremely seriously. So much so that I’ve been building a tool to help me optimize and monitor the sites that I build to make sure that I’m making an attempt to offer the best experience to those who visit them. If you’re into performant, accessible and SEO friendly sites, you might like it too! You can check it out at Optimize Toolset.

About

Hi, 👋, I’m Ryan Hefner  and I built this site for me, and you! The goal of this site was to provide an easy way for me to check the stats on my npm packages, both for prioritizing issues and updates, and to give me a little kick in the pants to keep up on stuff.

As I was building it, I realized that I was actually using the tool to build the tool, and figured I might as well put this out there and hopefully others will find it to be a fast and useful way to search and browse npm packages as I have.

If you’re interested in other things I’m working on, follow me on Twitter or check out the open source projects I’ve been publishing on GitHub.

I am also working on a Twitter bot for this site to tweet the most popular, newest, random packages from npm. Please follow that account now and it will start sending out packages soon–ish.

Open Software & Tools

This site wouldn’t be possible without the immense generosity and tireless efforts from the people who make contributions to the world and share their work via open source initiatives. Thank you 🙏

© 2025 – Pkg Stats / Ryan Hefner

node-red-contrib-coqui-stt

v0.1.1

Published

a node-red integration of the coqui stt component

Downloads

6

Readme

node-red-contrib-coqui-stt

a node-red node to perform speech to text inference using coqui stt

This suite of nodes uses the official coqui-stt node.js client cpu implementation. So just install the node from the palette or your node-red folder (normally ~/.node-red) with:

npm install node-red-contrib-coqui-stt

or directly from the repository with:

npm install johanneskropf/node-red-contrib-coqui-stt

(needs git installed) and coqui-stt will be automatically installed as a dependency.

Coqui-Wav Node

Basic Usage

To do speech to text inference you need to download a model(tflite) and a corresponding scorer file. For example the official english or chinese model can be found on the release page. You need to enter the path to both the model and the scorer in the nodes config. To do inference then send a wav buffer (16000Hz, 16bit, mono) to the nodes input in the configured msg input property. You will receive the transcription, input length and inference time as an object in the msg.payload or in your configured output property. If you want to do more accurate and quicker transcriptions of a limited vocabulary and sentences set you will need to train your own scorer file. Documentation on how to do this can be found in the coqui readme. For a list of some of the other available pre-trained models for different languages have a look on the coqui website in the model section.

Advanced

The node exposes a number other settings from the coqui-stt node.js api that can be used for changing the behaviour and tuning transcription speed and or results.

beam width

You can override the default beam width setting. The beamwidth influences how many options the deepspeech ctc beam search decoder explores while transcribing the audio. The higher the accuracy but the slower the transcription will become and vice versa. There is also a point of diminishing returns if set too high. The default is 512. Set lower for faster results or higher to see if a higher accuracy can be achieved with the model and scorer used.

lm alpha & beta

Each scorer comes with default lm alpha (language model weight) and lm beta (word insertion penalty) values. You can override those values if you have a reason. They are differnt for each scorer and the defaults are normally fairly well optimized.

disable external scorer

You can disable the use of an external scorer. This will give you the pure letter based predictions coming from the accoustic model. They will most likely not be very accurate and slower.

hotwords

This is a recent feature added to coqui-stt that allows you to increase the likelihood of certain words to appear in the transcription. Each hotword is accompanied by a boost value (between -100 and 100, negative values decrease the likelihood of appearance). A hotword should be a single word with no space that is part of the vocabulary used for the scorer. Most likely you wont need a value bigger than 10 to have a sufficient increase in propability for a single word. (using higher value can lead to detrimental results) You can add hotwords to the coqui nodes in node-red by sending an array of objects containing the words and boost values to the nodes configured input property:

[{"word":"car","boost":10},{"word":"dog","boost":15},{"word":"bird","boost":5}]

To clear the hotwords from the node send an empty array to it.

Coqui-Stream Node

Usage

For the stream node the same prerequisites as for the wav node apply. It also shares the same settings and advanced settings as described above. It differs in the following points:

  • this node accepts a stream of raw pcm audio buffers as its input. It will do the inference as the audio arrives. As soon as the audio stream stops it will send the transcription result in the configured msg property.
  • the stream needs to have the following format for most available models:
    • 16bit
    • little endian
    • signed-integer
    • 16000hz
    • mono
  • there is a timeout in milliseconds to consider the audio stream as stopped if no new buffers arrive that can be configured in the nodes settings. Setting the timeout to 0 will result in no stopping even if the audio stream stops.
  • in addition the stream inference node supports the following control messages send in the configured input property:
    • a string of stop in the configured message property will stop the running inference without returning a result
    • a string of stop_result in the configured message property will stop the running inference and return the final inference result
    • a string of intermediate in the configured message property will return an intermediate result with the transcription up to that point