@grapelaw/token-bucket
v1.0.2
Published
This package provides a simple and efficient implementation of the token bucket algorithm for rate limiting. It helps you control the frequency of actions (such as API requests) within a specified time frame.
Downloads
611
Maintainers
Readme
Token Bucket
This package provides a simple and efficient implementation of the token bucket algorithm for rate limiting. It helps you control the frequency of actions (such as API requests) within a specified time frame.
The token bucket algorithm provides a simple mechanism to control how many actions (e.g., API requests) can be performed over a given time window. The bucket accumulates tokens at a steady rate which you can fine-tune and stop it for a duration, and each action consumes a token. If no tokens are available, the action can either fail or wait for tokens to refill.
Features
- Asynchronous Consumption: Supports waiting for tokens asynchronously when they are unavailable.
- Customizable Refill Rates: Allows fine-tuning of bucket capacity, refill rate, and time windows.
- Force Stop and Refill: Can force a bucket to wait and stop refilling for a specified duration.
- Verbose Mode: Logs the state of the token bucket for debugging purposes.
Getting Started
Simple Usage
import { TokenBucket } from 'token-bucket';
const tokenBucket = new TokenBucket({
capacity: 10,
fillPerWindow: 1,
windowInMs: 1000,
});
// Simple Usage: Returns true only when a token is available right away
if (tokenBucket.consume()) {
// Your logic IF the required tokens are available right away.
}
// Async Usage: Wait until a token is available
if (await tokenBucket.consumeAsync()) {
// Your logic WHEN required tokens are consumed successfully
}
// Async Usage 2: Wait until a token is available
while (!await tokenBucket.consumeAsync()); // Don't forget to add a semicolon after the while loop
// Your logic here, which will run AFTER the token is available and consumed.
Usage with Verbose Logging
import { TokenBucket } from 'token-bucket';
const tokenBucket = new TokenBucket({
capacity: 20,
fillPerWindow: 5,
windowInMs: 2000,
}, true);
if (await tokenBucket.consumeAsync()) {
// Your async logic when a token is consumed successfully
}
// Logs the following
// * When there are not enough tokens, it will log a message, which also tells when will the tokens be available
// * Every 10 seconds, the remaining tokens
// * Forced Stop started-ended
Singleton Recommendation
import { TokenBucket } from 'token-bucket';
export const tokenBucket = new TokenBucket({
capacity: 20,
fillPerWindow: 5,
windowInMs: 2000,
}, true);
// For utilizing the same api / resource, using a single instance of tokenBucket is recommended throughout the application is recommended.
Configuration Options
TokenBucketOptions
capacity
: The maximum number of tokens in the bucket.fillPerWindow
: The number of tokens added to the bucket per window.windowInMs
: The size of the window in milliseconds.initialTokens
: The initial number of tokens in the bucket. Defaults to the capacity if not provided.
import { TokenBucket } from 'token-bucket';
const tokenBucket = new TokenBucket({
capacity: 20,
fillPerWindow: 2,
windowInMs: 1500,
initialTokens: 40,
});
Methods
consume(amount: number = 1): boolean
Consumes the specified number of tokens from the bucket. Returns true if the tokens were successfully consumed, otherwise false.
consumeAsync(amount: number = 1): Promise<boolean>
Asynchronously consumes the specified number of tokens from the bucket. Returns a promise that resolves to true if the tokens were successfully consumed, otherwise false.
start(): TokenBucket
Starts the internal timer for refilling tokens. Returns the TokenBucket instance to allow chaining.
stop(): void
Stops the internal timer for refilling tokens.
forceWaitUntilMilisecondsPassed(delayInMs: number): void
Forces the bucket to wait until the specified amount of milliseconds have passed before refilling. The intended use of this method is aligning with the rate limit of the API in case your TokenBucket is providing more tokens than it should (i.e: if you are using shared rate limit within different places).
Authors
Nurbaki Kasikci - GitHub - Twitter
Contribution
We welcome contributions to improve this package and encourage users to submit bug reports, feature requests, or any other contributions that can enhance the project. Please follow the guidelines below to contribute:
- Report Issues: If you encounter any issues or have suggestions for improvements, please open an issue on GitHub
- Pull Requests: You are welcome to submit Pull Requests (PRs) for bug fixes or new features. Make sure to follow the established coding conventions and explain the purpose of your changes.