@event-driven-io/pongo
v0.16.3
Published
Pongo - Mongo with strong consistency on top of Postgres
Downloads
385
Readme
Pongo
Pongo - Mongo but on Postgres and with strong consistency benefits.
Getting Started
Install Pongo as an npm module and save it to your package.json:
npm install @event-driven-io/pongo
Read also introduction article on my blog for more context.
Example
You can use Pongo syntax with explicit typing about supported syntax:
import { pongoClient, ObjectId } from "@event-driven-io/pongo";
type User = { name: string; age: number };
const connectionString =
"postgresql://dbuser:[email protected]:3211/mydb";
const pongo = pongoClient(connectionString);
const pongoDb = pongo.db();
const users = pongoDb.collection<User>("users");
const roger = { name: "Roger", age: 30 };
const anita = { name: "Anita", age: 25 };
const cruella = { _id: ObjectId(), name: "Cruella", age: 40 };
// Inserting
await users.insertOne(roger);
await users.insertOne(cruella);
const { insertedId } = await users.insertOne(anita);
const anitaId = insertedId;
// Updating
await users.updateOne({ _id: anitaId }, { $set: { age: 31 } });
// Deleting
await users.deleteOne({ _id: cruella._id });
// Finding by Id
const anitaFromDb = await users.findOne({ _id: anitaId });
// Finding more
const usersFromDb = await users.find({ age: { $lt: 40 } });
Or use MongoDB compliant shim:
import { MongoClient, ObjectId } from "@event-driven-io/pongo/shim";
type User = { name: string; age: number };
const connectionString =
"postgresql://dbuser:[email protected]:3211/mydb";
const pongoClient = new MongoClient(postgresConnectionString);
const pongoDb = pongoClient.db();
const users = pongoDb.collection<User>("users");
const roger = { name: "Roger", age: 30 };
const anita = { name: "Anita", age: 25 };
const cruella = { _id: ObjectId(), name: "Cruella", age: 40 };
// Inserting
await users.insertOne(roger);
await users.insertOne(cruella);
const { insertedId } = await users.insertOne(anita);
const anitaId = insertedId;
// Updating
await users.updateOne({ _id: anitaId }, { $set: { age: 31 } });
// Deleting
await users.deleteOne({ _id: cruella._id });
// Finding by Id
const anitaFromDb = await users.findOne({ _id: anitaId });
// Finding more
const usersFromDb = await users.find({ age: { $lt: 40 } }).toArray();
How does it work?
Pongo treats PostgreSQL as a Document Database benefiting from JSONB support. Unlike the plain text storage of the traditional JSON type, JSONB stores JSON data in a binary format. This simple change brings significant advantages in terms of performance and storage efficiency.
Pongo uses the following table structure for storing collections:
CREATE TABLE IF NOT EXISTS "YourCollectionName" (
_id TEXT PRIMARY KEY,
data JSONB NOT NULL,
metadata JSONB NOT NULL DEFAULT '{}',
_version BIGINT NOT NULL DEFAULT 1,
_partition TEXT NOT NULL DEFAULT 'png_global',
_archived BOOLEAN NOT NULL DEFAULT FALSE,
_created TIMESTAMPTZ NOT NULL DEFAULT now(),
_updated TIMESTAMPTZ NOT NULL DEFAULT now()
)
Essentially Pongo takes MongoDB api and translates it to the native PostgreSQL queries. It is a similar concept to Marten, FerretDB and AWS DocumentDB.
E.g. the MongoDB update syntax:
const users = pongoDb.collection<User>("users");
await users.updateOne({ _id: someId }, { $push: { tags: "character" } });
will be translated to:
UPDATE "users"
SET data = jsonb_set(data, '{tags}', (COALESCE(data->'tags', '[]'::jsonb) || to_jsonb('character')))
WHERE _id = '137ef052-e41c-428b-b606-1c8070a47eda';
Or for query:
const result = await users
.find({ "address.history": { $elemMatch: { street: "Elm St" } } })
.toArray();
will result in:
SELECT data
FROM "users"
WHERE jsonb_path_exists(
data,
'$.address.history[*] ? (@.street == "Elm St")'
);
Why Pongo?
MongoDB is a decent database, yet it has issues around ACID-complaince and licensing, which can cause hardship for project scenarios and organisation policies.
Pongo brings the PostgreSQL shape-shifting capabilities to:
- benefit from strong consistency by using battle-tested and widely used PostgreSQL ACID-compliant database,
- easier integration with other parts of your system using PostgreSQL,
- reuse your muscle memory from MongoDB using compatible API. It will allow easier migration of existing projects,
- cut boilerplate and easier nested data management than traditional relational tables,
- operate easier than crafting native PostgreSQL JSON queries. They're powerful but not the most accessible,
- get performance boost with JSONB indexing capabilities,
- benefit from PostgreSQL advanced capabilities like partitioning, logical replication and other PostgreSQL superpowers
- seamless integration with Cloud RDSes and solutions like Supabase, Vercel Postgres, YugabyteDB, CockroachDB, .
Watch also more in:
Storage
The binary format of PostgreSQL JSONB means that data is pre-parsed, allowing faster read and write operations than text-based JSON. You don't have to re-parse the data every time you query it, which saves processing time and improves overall performance. Additionally, JSONB supports advanced indexing options like GIN and GiST indexes, making searches within JSONB documents much quicker and more efficient.
Moreover, JSONB retains the flexibility of storing semi-structured data while allowing you to use PostgreSQL's robust querying capabilities. You can perform complex queries, joins, and transactions with JSONB data, just as you can with regular relational data.
Contrary to common belief, JSON document data is structured. JSON has structure, but it is not enforced for each document. We can easily extend the schema for our documents, even for specific ones, by adding new fields. We should also not fail if a field we expect to exist doesn't.
This flexibility, performance, and consistency combination makes PostgreSQL with JSONB a powerful tool. There are benchmarks showing that it can be even faster than MongoDB.
Check more in:
- JSON Types Documentation
- JSON Functions and Operators
- PostgreSQL, JSONB and GIN Indexes by
- MongoDB vs PostgreSQL JSONB Benchmark
- How to JSON in PostgreSQL
- Just Use Postgres for Everything
Is Pongo an ORM?
It's not.
It's focused on effective handling of the document data specifics. Node.js ORMs have capabilities to handle JSONB, e.g. DrizzleORM has good support for that for basic operations.
Yet, they all have limited querying capabilities. Usually for advanced ones you need to fallback to JSONPath or JSONB functions (so raw SQL). As you saw above, this syntax is not super pleasant to deal with. That's why Pongo aims to do it for you.
How is it different than FerretDB?
FerretDB plugs into the native MongoDB protocol, which allows it to be used as MongoDB and connect to tools like Mongo UI, etc. Yet, it requires running a proxy.
Pongo operates on a different layer, translating the MongoDB API directly into SQL in the library code. This can allow easier serverless integrations, such as sharing a connection pool with other PostgreSQL-based tools, etc. Of course, it won't allow using native tools based on the MongoDB network protocol.
Pongo's goal is not to replace Mongo but to reuse its muscle memory and bring the PostgreSQL capabilities and superpowers into the Node.js land.
Is it production ready?
What's there is safe to use, but it's far from being 100% compliant with MongoDB. Pongo is a fresh project, so some stuff can be missing.
Contribution
Pongo is a community project, so once you find something missing or not working, we encourage you to send us a GH issue or Pull Request extending the support or test coverage! Check also Contributing guide
If you think something is missing or want to get some features faster, I'm happy to take sponsoring to prioritise it. Feel free to contact me - we'll find a way to help you!
Code of Conduct
This project has adopted the code of conduct defined by the Contributor Covenant to clarify expected behavior in our community.