npm package discovery and stats viewer.

Discover Tips

  • General search

    [free text search, go nuts!]

  • Package details

    pkg:[package-name]

  • User packages

    @[username]

Sponsor

Optimize Toolset

I’ve always been into building performant and accessible sites, but lately I’ve been taking it extremely seriously. So much so that I’ve been building a tool to help me optimize and monitor the sites that I build to make sure that I’m making an attempt to offer the best experience to those who visit them. If you’re into performant, accessible and SEO friendly sites, you might like it too! You can check it out at Optimize Toolset.

About

Hi, 👋, I’m Ryan Hefner  and I built this site for me, and you! The goal of this site was to provide an easy way for me to check the stats on my npm packages, both for prioritizing issues and updates, and to give me a little kick in the pants to keep up on stuff.

As I was building it, I realized that I was actually using the tool to build the tool, and figured I might as well put this out there and hopefully others will find it to be a fast and useful way to search and browse npm packages as I have.

If you’re interested in other things I’m working on, follow me on Twitter or check out the open source projects I’ve been publishing on GitHub.

I am also working on a Twitter bot for this site to tweet the most popular, newest, random packages from npm. Please follow that account now and it will start sending out packages soon–ish.

Open Software & Tools

This site wouldn’t be possible without the immense generosity and tireless efforts from the people who make contributions to the world and share their work via open source initiatives. Thank you 🙏

© 2024 – Pkg Stats / Ryan Hefner

@posthog/plugin-server

v1.10.7

Published

PostHog Plugin Server

Downloads

5,133

Readme

PostHog Plugin Server

npm package MIT License

This service takes care of processing events with plugins and more.

Get started

Let's get you developing the plugin server in no time:

  1. Have virtual environment from the main PostHog repo active.

  2. Install dependencies and prepare for takeoff by running command yarn.

  3. Start a development instance of PostHog - instructions here. After all, this is the PostHog Plugin Server, and it works in conjuction with the main server. To avoid interference, disable the plugin server there with setting the PLUGIN_SERVER_IDLE env variable before running. PLUGIN_SERVER_IDLE=true ./bin/start

  4. Make sure that the plugin server is configured correctly (see Configuration). Two settings that you MUST get right are DATABASE_URL and REDIS_URL - they need to be identical between the plugin server and the main server.

  5. If developing the enterprise Kafka + ClickHouse pipeline, set KAFKA_ENABLED to true and provide KAFKA_HOSTS plus CLICKHOUSE_HOST, CLICKHOUSE_DATABASE, CLICKHOUSE_USER, andCLICKHOUSE_PASSWORD.

    Otherwise if developing the basic Redis + Postgres pipeline, skip ahead.

  6. Start the plugin server in autoreload mode with yarn start, or in compiled mode with yarn build && yarn start:dist, and develop away!

  7. To run migrations for the test, run yarn setup:test:postgres or setup:test:clickhouse. Run Postgres pipeline tests with yarn test:postgres:{1,2}. Run ClickHouse pipeline tests with yarn test:clickhouse:{1,2}. Run benchmarks with yarn benchmark.

Alternative modes

This program's main mode of operation is processing PostHog events, but there are also a few alternative utility ones. Each one does a single thing. They are listed in the table below, in order of precedence.

| Name | Description | CLI flags | | ----------- | ---------------------------------------------------------- | ----------------- | | Help | Show plugin server configuration options | -h, --help | | Version | Only show currently running plugin server version | -v, --version | | Healthcheck | Check plugin server health and exit with 0 or 1 | --healthcheck | | Migrate | Migrate Graphile job queue | --migrate | | Idle | Start server in a completely idle, non-processing mode | --idle |

Configuration

There's a multitude of settings you can use to control the plugin server. Use them as environment variables.

| Name | Description | Default value | | ----------------------------- | -------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- | ------------------------------------- | | DATABASE_URL | Postgres database URL | 'postgres://localhost:5432/posthog' | | REDIS_URL | Redis store URL | 'redis://localhost' | | BASE_DIR | base path for resolving local plugins | '.' | | WORKER_CONCURRENCY | number of concurrent worker threads | 0 – all cores | | TASKS_PER_WORKER | number of parallel tasks per worker thread | 10 | | REDIS_POOL_MIN_SIZE | minimum number of Redis connections to use per thread | 1 | | REDIS_POOL_MAX_SIZE | maximum number of Redis connections to use per thread | 3 | | SCHEDULE_LOCK_TTL | how many seconds to hold the lock for the schedule | 60 | | CELERY_DEFAULT_QUEUE | Celery outgoing queue | 'celery' | | PLUGINS_CELERY_QUEUE | Celery incoming queue | 'posthog-plugins' | | PLUGINS_RELOAD_PUBSUB_CHANNEL | Redis channel for reload events | 'reload-plugins' | | CLICKHOUSE_HOST | ClickHouse host | 'localhost' | | CLICKHOUSE_DATABASE | ClickHouse database | 'default' | | CLICKHOUSE_USER | ClickHouse username | 'default' | | CLICKHOUSE_PASSWORD | ClickHouse password | null | | CLICKHOUSE_CA | ClickHouse CA certs | null | | CLICKHOUSE_SECURE | whether to secure ClickHouse connection | false | | KAFKA_ENABLED | use Kafka instead of Celery to ingest events | false | | KAFKA_HOSTS | comma-delimited Kafka hosts | null | | KAFKA_CONSUMPTION_TOPIC | Kafka incoming events topic | 'events_plugin_ingestion' | | KAFKA_CLIENT_CERT_B64 | Kafka certificate in Base64 | null | | KAFKA_CLIENT_CERT_KEY_B64 | Kafka certificate key in Base64 | null | | KAFKA_TRUSTED_CERT_B64 | Kafka trusted CA in Base64 | null | | KAFKA_PRODUCER_MAX_QUEUE_SIZE | Kafka producer batch max size before flushing | 20 | | KAFKA_FLUSH_FREQUENCY_MS | Kafka producer batch max duration before flushing | 500 | | KAFKA_MAX_MESSAGE_BATCH_SIZE | Kafka producer batch max size in bytes before flushing | 900000 | | LOG_LEVEL | minimum log level | 'info' | | SENTRY_DSN | Sentry ingestion URL | null | | STATSD_HOST | StatsD host - integration disabled if this is not provided | null | | STATSD_PORT | StatsD port | 8125 | | STATSD_PREFIX | StatsD prefix | 'plugin-server.' | | DISABLE_MMDB | whether to disable MMDB IP location capabilities | false | | INTERNAL_MMDB_SERVER_PORT | port of the internal server used for IP location (0 means random) | 0 | | DISTINCT_ID_LRU_SIZE | size of persons distinct ID LRU cache | 10000 | | PLUGIN_SERVER_IDLE | whether to disengage the plugin server, e.g. for development | false | | CAPTURE_INTERNAL_METRICS | whether to capture internal metrics for posthog in posthog | false | | PISCINA_USE_ATOMICS | corresponds to the piscina useAtomics config option (https://github.com/piscinajs/piscina#constructor-new-piscinaoptions) | true | | PISCINA_ATOMICS_TIMEOUT | (advanced) corresponds to the length of time (in ms) a piscina worker should block for when looking for tasks - instances with high volumes (100+ events/sec) might benefit from setting this to a lower value | 5000 |

Releasing a new version

Just bump up version in package.json on the main branch and the new version will be published automatically, with a matching PR in the main PostHog repo created.

It's advised to use bump patch/minor/major label on PRs - that way the above will be done automatically when the PR is merged.

Courtesy of GitHub Actions.

Walkthrough

The story begins with pluginServer.ts -> startPluginServer, which is the main thread of the plugin server.

This main thread spawns WORKER_CONCURRENCY worker threads, managed using Piscina. Each worker thread runs TASKS_PER_WORKER tasks (concurrentTasksPerWorker).

Main thread

Let's talk about the main thread first. This has:

  1. pubSub – Redis powered pub-sub mechanism for reloading plugins whenever a message is published by the main PostHog app.

  2. hub – Handler of connections to required DBs and queues (ClickHouse, Kafka, Postgres, Redis), holds loaded plugins. Created via hub.ts -> createHub. Every thread has its own instance.

  3. piscina – Manager of tasks delegated to threads. makePiscina creates the manager, while createWorker creates the worker threads.

  4. scheduleControl – Controller of scheduled jobs. Responsible for adding Piscina tasks for scheduled jobs, when the time comes. The schedule information makes it into the controller when plugin VMs are created.

    Scheduled tasks are controlled with Redlock (redis-based distributed lock), and run on only one plugin server instance in the entire cluster.

  5. jobQueueConsumer – The internal job queue consumer. This enables retries, scheduling jobs in the future (once) (Note: this is the difference between scheduleControl and this internal jobQueue). While scheduleControl is triggered via runEveryMinute, runEveryHour tasks, the jobQueueConsumer deals with meta.jobs.doX(event).runAt(new Date()).

    Jobs are enqueued by job-queue-manager.ts, which is backed by Postgres-based Graphile-worker (graphile-queue.ts).

  6. queue – Event ingestion queue. This is a Celery (backed by Redis) or Kafka queue, depending on the setup (EE/Cloud is Kafka due to high volume). These are consumed by the queue above, and sent off to the Piscina workers (src/main/ingestion-queues/queue.ts -> ingestEvent). Since all of the actual ingestion happens inside worker threads, you'll find the specific ingestion code there (src/worker/ingestion/ingest-event.ts). There the data is saved into Postgres (and ClickHouse via Kafka on EE/Cloud).

    It's also a good idea to see the producer side of this ingestion queue, which comes from Posthog/posthog/api/capture.py. The plugin server gets the process_event_with_plugins Celery task from there, in the Postgres pipeline. The ClickHouse via Kafka pipeline gets the data by way of Kafka topic events_plugin_ingestion.

  7. mmdbServer – TCP server, which works as an interface between the GeoIP MMDB data reader located in main thread memory and plugins ran in worker threads of the same plugin server instance. This way the GeoIP reader is only loaded in one thread and can be used in all. Additionally this mechanism ensures that mmdbServer is ready before ingestion is started (database downloaded from http-mmdb and read), and keeps the database up to date in the background.

Worker threads

This begins with worker.ts and createWorker().

hub is the same setup as in the main thread.

New functions called here are:

  1. setupPlugins – Loads plugins and prepares them for lazy VM initialization.

  2. createTaskRunner – Creates a Piscina task runner that allows to operate on plugin VMs.

Note: An organization_id is tied to a company and its installed plugins, a team_id is tied to a project and its plugin configs (enabled/disabled+extra config).

Questions?

Join our Slack community. 🦔