npm package discovery and stats viewer.

Discover Tips

  • General search

    [free text search, go nuts!]

  • Package details

    pkg:[package-name]

  • User packages

    @[username]

Sponsor

Optimize Toolset

I’ve always been into building performant and accessible sites, but lately I’ve been taking it extremely seriously. So much so that I’ve been building a tool to help me optimize and monitor the sites that I build to make sure that I’m making an attempt to offer the best experience to those who visit them. If you’re into performant, accessible and SEO friendly sites, you might like it too! You can check it out at Optimize Toolset.

About

Hi, 👋, I’m Ryan Hefner  and I built this site for me, and you! The goal of this site was to provide an easy way for me to check the stats on my npm packages, both for prioritizing issues and updates, and to give me a little kick in the pants to keep up on stuff.

As I was building it, I realized that I was actually using the tool to build the tool, and figured I might as well put this out there and hopefully others will find it to be a fast and useful way to search and browse npm packages as I have.

If you’re interested in other things I’m working on, follow me on Twitter or check out the open source projects I’ve been publishing on GitHub.

I am also working on a Twitter bot for this site to tweet the most popular, newest, random packages from npm. Please follow that account now and it will start sending out packages soon–ish.

Open Software & Tools

This site wouldn’t be possible without the immense generosity and tireless efforts from the people who make contributions to the world and share their work via open source initiatives. Thank you 🙏

© 2024 – Pkg Stats / Ryan Hefner

@sourceloop/bpmn-service

v14.0.1

Published

BPMN microservice.

Downloads

572

Readme

@sourceloop/bpmn-service

LoopBack

npm

node-current (scoped)

npm (prod) dependency version (scoped)

Overview

A Microservice for handling BPMN workflows using engines like Camunda.

NOTE: The microservice currently works with only one workflow definition for a single diagram. It provides -

  • Deployment and management of Workflows in a BPMN engine.
  • Process Versioning.
  • Executing Workflow using an endpoint, validate inputs before starting an execution.

Sandbox Example

The sandbox example provides a use of this microservice with the Camunda BPMN engine.

It uses camunda-external-task-client-js to implement external task workers, that are triggered on executing a workflow. The flow of control is shown in the diagram below -

Execute Workflow

Installation

npm i @sourceloop/bpmn-service

Usage

  • Create a new Loopback4 Application (If you don't have one already) lb4 testapp
  • Install the bpmn service npm i @sourceloop/bpmn-service
  • Set the environment variables.
  • Run the migrations.
  • Bind the BPMN Config to WorkflowServiceBindings.Config key-
    this.bind(WorkflowServiceBindings.Config).to({
      useCustomSequence: true,
      workflowEngineBaseUrl: process.env.CAMUNDA_URL, // url for the rest engine in case of Camunda
    });
  • Implement WorkflowProvider (refer this) and bind it to WorkflowServiceBindings.WorkflowManager key -
    this.bind(WorkflowServiceBindings.WorkflowManager).toProvider(
      WorkflowProvider,
    );
  • Add the WorkflowServiceComponent to your Loopback4 Application (in application.ts).
    // import WorkflowServiceComponent
    import {WorkflowServiceComponent} from '@sourceloop/bpmn-service';
    // add Component for WorkflowService
    this.component(WorkflowServiceComponent);
  • Set up a Loopback4 Datasource with dataSourceName property set to WorkflowCacheSourceName. You can see an example datasource here.
  • Start the application npm start

Using with Sequelize

This service supports Sequelize as the underlying ORM using @loopback/sequelize extension. And in order to use it, you'll need to do following changes.

  • To use Sequelize in your application, add following to application.ts along with other config specific to the service:

    this.bind(WorkflowServiceBindings.Config).to({
      useCustomSequence: false,
      useSequelize: true,
    });
  • Use the SequelizeDataSource in your audit datasource as the parent class. Refer this for more.

Setting up a DataSource

Here is a sample Implementation DataSource implementation using environment variables and PostgreSQL as the data source.

import {inject, lifeCycleObserver, LifeCycleObserver} from '@loopback/core';
import {juggler} from '@loopback/repository';
import {WorkflowCacheSourceName} from '../types';

const config = {
  name: WorkflowCacheSourceName,
  connector: 'postgresql',
  host: process.env.DB_HOST,
  port: process.env.DB_PORT,
  user: process.env.DB_USER,
  password: process.env.DB_PASSWORD,
  database: process.env.DB_DATABASE,
  schema: process.env.DB_SCHEMA,
};

@lifeCycleObserver('datasource')
export class BpmnDbDataSource
  extends juggler.DataSource
  implements LifeCycleObserver
{
  static dataSourceName = WorkflowCacheSourceName;
  static readonly defaultConfig = config;

  constructor(
    @inject('datasources.config.BpmnDb', {optional: true})
    dsConfig: object = config,
  ) {
    super(dsConfig);
  }
}

Asymmetric Token Signing and Verification

If you are using asymmetric token signing and verification, you need to create a datasource for auth database. Example datasource file for auth:-

import {inject, lifeCycleObserver, LifeCycleObserver} from '@loopback/core';
import {juggler} from '@loopback/repository';
import {AuthDbSourceName} from '@sourceloop/core';
const DEFAULT_MAX_CONNECTIONS = 25;
const DEFAULT_DB_IDLE_TIMEOUT_MILLIS = 60000;
const DEFAULT_DB_CONNECTION_TIMEOUT_MILLIS = 2000;

const config = {
  name: 'auth',
  connector: 'postgresql',
  host: process.env.DB_HOST,
  port: process.env.DB_PORT,
  user: process.env.DB_USER,
  schema: process.env.DB_SCHEMA,
  password: process.env.DB_PASSWORD,
  database: process.env.AUTH_DB,
};

// Observe application's life cycle to disconnect the datasource when
// application is stopped. This allows the application to be shut down
// gracefully. The `stop()` method is inherited from `juggler.DataSource`.
// Learn more at https://loopback.io/doc/en/lb4/Life-cycle.html
@lifeCycleObserver('datasource')
export class AuthDataSource
  extends juggler.DataSource
  implements LifeCycleObserver
{
  static dataSourceName = AuthDbSourceName;

  static readonly defaultConfig = config;

  constructor(
    @inject('datasources.config.auth', {optional: true})
    dsConfig: object = config,
  ) {
    if (!!+(process.env.ENABLE_DB_CONNECTION_POOLING ?? 0)) {
      const dbPool = {
        max: +(process.env.DB_MAX_CONNECTIONS ?? DEFAULT_MAX_CONNECTIONS),
        idleTimeoutMillis: +(
          process.env.DB_IDLE_TIMEOUT_MILLIS ?? DEFAULT_DB_IDLE_TIMEOUT_MILLIS
        ),
        connectionTimeoutMillis: +(
          process.env.DB_CONNECTION_TIMEOUT_MILLIS ??
          DEFAULT_DB_CONNECTION_TIMEOUT_MILLIS
        ),
      };

      dsConfig = {...dsConfig, ...dbPool};
    }

    super(dsConfig);
  }
}

Environment Variables

The service comes with a default DataSource using PostgreSQL, if you intend to use this, you have to provide the following variables in the environment -

| Name | Required | Default Value | Description | | ------------- | -------- | ------------- | ---------------------------------------------------------------------------------------------------------------------------------- | | DB_HOST | Y | | Hostname for the database server. | | DB_PORT | Y | | Port for the database server. | | DB_USER | Y | | User for the database. | | DB_PASSWORD | Y | | Password for the database user. | | DB_DATABASE | Y | | Database to connect to on the database server. | | DB_SCHEMA | Y | | Database schema used for the data source. In PostgreSQL, this will be public unless a schema is made explicitly for the service. |

Setting up a DataSource

A sample implementation of a DataSource using environment variables and PostgreSQL is included with the service, you can provide your own using the BpmnDbSourceName variable. Implementation of the sample DataSource can be seen here.

Database Schema

bpmndb

Providers

  • BPMNProvider

    To use the services, you need to implement a provider and bind it to the BPMNBindings.BPMNProvider key. The provider returns a value containing the 5 methods - getWorkflowById, startWorkflow, createWorkflow, updateWorkflow and deleteWorkflowById. These methods are responsible for performing their respective tasks in the workflow engine. Here is the default implementation of this provider -

    import {bind, /* inject, */ BindingScope, Provider} from '@loopback/core';
    import {HttpErrors} from '@loopback/rest';
    import {WorflowManager} from '../types';
    
    @bind({scope: BindingScope.TRANSIENT})
    export class WorkflowProvider implements Provider<WorflowManager> {
      value() {
        return {
          getWorkflowById: async () => {
            throw new HttpErrors.BadRequest(
              'getWorkflowId function not implemented',
            );
          },
          startWorkflow: async () => {
            throw new HttpErrors.BadRequest(
              'startWorkflow function not implemented',
            );
          },
          createWorkflow: async () => {
            throw new HttpErrors.BadRequest(
              'createWorkflow function not implemented',
            );
          },
          updateWorkflow: async () => {
            throw new HttpErrors.BadRequest(
              'updateWorkflow function not implemented',
            );
          },
          deleteWorkflowById: async () => {
            throw new HttpErrors.BadRequest(
              'deleteWorkflowById function not implemented',
            );
          },
        };
      }
    }
  • WorkerImplementationProvider

    Your workers are automatically initiated once a workflow is executed, to provide the implementation details of workers, you need to give an implementation template of one such worker using the WorkflowServiceBindings.WorkerImplementationFunction, a default implementation is provided here. You also need to register individual worker commands using the WorkflowServiceBindings.RegisterWorkerFunction function;

  • ExecutionInputValidationProvider

    If you need to validate the inputs of a workflow execution, you can bind a custom validation provider using WorkflowServiceBindings.ExecutionInputValidatorFn key. The microservice comes with a default implementation using AJV.

Migrations

The migrations required for this service are processed during the installation automatically if you set the WORKFLOW_MIGRATION or SOURCELOOP_MIGRATION env variable. The migrations use db-migrate with db-migrate-pg driver for migrations, so you will have to install these packages to use auto-migration. Please note that if you are using some pre-existing migrations or databases, they may be affected. In such a scenario, it is advised that you copy the migration files in your project root, using the WORKFLOW_MIGRATION_COPY or SOURCELOOP_MIGRATION_COPY env variables. You can customize or cherry-pick the migrations in the copied files according to your specific requirements and then apply them to the DB.

This migration script supports both MySQL and PostgreSQL databases, controlled by environment variables. By setting MYSQL_MIGRATION to 'true', the script runs migrations using MySQL configuration files; otherwise, it defaults to PostgreSQL. .

This project includes no migrations to seed your BPMN engine. If you are using Camunda BPM Run, you can use either the resources folder to seed a model, or you can config it to use a custom DB where you can seed your data. The steps to config Platform Run are given here.

Additionally, there is now an option to choose between SQL migration or PostgreSQL migration. NOTE: For @sourceloop/cli users, this choice can be specified during the scaffolding process by selecting the "type of datasource" option.

API Documentation

Common Headers

Authorization: Bearer where is a JWT token signed using JWT issuer and secret. Content-Type: application/json in the response and in request if the API method is NOT GET

Common Request path Parameters

{version}: Defines the API Version

Common Responses

200: Successful Response. Response body varies w.r.t API 401: Unauthorized: The JWT token is missing or invalid 403: Forbidden : Not allowed to execute the concerned API 404: Entity Not Found 400: Bad Request (Error message varies w.r.t API) 201: No content: Empty Response

API's Details

Visit the OpenAPI spec docs