@upstash/queue
v1.0.1
Published
A message queue based on Redis streams, without dependencies.
Downloads
187
Readme
Upstash Queue
[!NOTE]
This project is in the Experimental Stage. We declare this project experimental to set clear expectations for your usage. There could be known or unknown bugs, the API could evolve, or the project could be discontinued if it does not find community adoption. While we cannot provide professional support for experimental projects, we’d be happy to hear from you if you see value in this project!
A simple, fast, robust, stream-based message queue for Node.js, backed by Upstash Redis, inspired by AWS SQS.
- Simple: ~350 LOC, and single dependency.
- Lightweight: Under ~5kb zipped.
- Fast: Maximizes throughput by minimizing Redis and network overhead. Benchmarks well.
- Robust: Designed with concurrency and failure in mind; full code coverage.
import { Redis } from "@upstash/redis";
import { Queue } from "@upstash/queue";
const queue = new Queue({ redis: new Redis() });
await queue.sendMessage({ hello: "world1" });
const message1 = await queue.receiveMessage<{ hello: "world1" }>();
expect(message1?.body).toEqual({ hello: "world1" });
Introduction
@upstash/queue
is a Node.js library that provides a simple and efficient way to implement a message queue system using Redis streams. It offers features such as message sending, receiving, automatic message verification, and concurrency control. This library is particularly useful for building distributed systems and background job processing.
Why Upstash Queue?
Upstash Queue brings the simplicity and performance of Redis streams to Node.js developers, making it easy to integrate a robust message queue system into their applications. Whether you're working on distributed systems, background job processing, or other asynchronous workflows, Upstash Queue provides essential features such as message sending, receiving, automatic verification, and concurrency control.
Key Features:
- Efficiency: Leverages the speed and reliability of Redis streams for message handling.
- Simplicity: Dead simple implementation of distributed systems and background job processing with a clean and intuitive API.
- Concurrency Control: Easily manages concurrent message processing to optimize system performance.
- Automatic Verification: Benefits from built-in message verification mechanisms for reliable and secure communication.
Table of Contents
Installation
To start using Upstash Queue, install the library via npm:
npm install @upstash/queue
Getting Started
import Redis from "@upstash/redis";
import { Queue } from "@upstash/queue";
const redis = new Redis();
const queue = new Queue({ redis });
Sending a Message
const payload = { key: "value" };
const delayInSeconds = 0; // Set to 0 for immediate delivery.
const streamId = await queue.sendMessage(payload, delayInSeconds);
Receiving a Message
const pollingForNewMessages = 1000; // Set to 0 for non-blocking, otherwise, it will try to get a message then fail if none is available.
const receivedMessage = await queue.receiveMessage(pollingForNewMessages);
console.log("Received Message:", receivedMessage);
Verifying a Message
const streamId = "some_stream_id";
const verificationStatus = await queue.verifyMessage(streamId);
console.log("Verification Status:", verificationStatus);
Configuration Options
When initializing the Queue instance, you can provide various configuration options:
- redis: Redis client instance (required).
- queueName: Name of the Redis stream (default: "UpstashMQ:Queue").
- concurrencyLimit: Maximum concurrent message processing allowed (default: 1).
- autoVerify: Auto-verify received messages (default: true).
- consumerGroupName: Group that holds consumers when automatically created (default: "Messages").
- visibilityTimeout: Time until recently sent messages become visible to other consumers (default: 30 seconds).
Verifying a Message
const queueConfig = {
redis: new Redis(),
queueName: "MyCustomQueue",
concurrencyLimit: 2,
autoVerify: false,
consumerGroupName: "MyConsumers",
visibilityTimeout: 60000, // 1 minute.
};
const customQueue = new Queue(queueConfig);
Examples
FIFO Example
const queue = new Queue({ redis });
await queue.sendMessage({ hello: "world1" });
await delay(100);
await queue.sendMessage({ hello: "world2" });
await delay(100);
await queue.sendMessage({ hello: "world3" });
const message1 = await queue.receiveMessage<{ hello: "world1" }>(); // Logs out { hello: "world1" }.
const message2 = await queue.receiveMessage<{ hello: "world2" }>(); // Logs out { hello: "world2" }.
const message3 = await queue.receiveMessage<{ hello: "world3" }>(); // Logs out { hello: "world3" }.
Sending Message with Delay then Poll
const fakeValue = randomValue();
const queueName = "app-logs";
const producer = new Queue({ redis, queueName });
const consumer = new Queue({
redis: new Redis(),
queueName,
});
await producer.sendMessage(
{
dev: fakeValue,
},
2000
);
const receiveMessageRes = await consumer.receiveMessage<{
dev: string;
}>(5000);
Manual Verification
const queue = new Queue({ redis, autoVerify: false });
await queue.sendMessage({ hello: "world" });
const message = await queue.receiveMessage<{ hello: "world" }>(); // Logs out { hello: "world" }.
if (message) {
await queue.verifyMessage(message.streamId); //Logs out "VERIFIED" or "NOT VERIFIED".
}
Concurrent Message Processing/Consuming
If concurrencyLimit
is not set, one of the receiveMessage()
will throw. You need to explicitly set the concurrencyLimit. Default is 1.
const queue = new Queue({
redis: new Redis(),
queueName: randomValue(),
concurrencyLimit: 2,
});
await queue.sendMessage({
dev: randomValue(),
});
await queue.sendMessage({
dev: randomValue(),
});
await Promise.all([queue.receiveMessage(), queue.receiveMessage()]);