zubairov-sailor-nodejs
v2.2.3-dev.1
Published
The official elastic.io library for bootstrapping and executing for Node.js connectors
Downloads
3
Readme
elasticio-sailor-nodejs
The official elastic.io library for bootstrapping and executing for Node.js connectors.
elasticio-sailor-nodejs
is a required dependency for all components build for elastic.io platform in Node.js. Add the dependency in the package.json
file in the following way:
"dependencies": {
"q": "^1.4.1",
"elasticio-sailor-nodejs": "^2.2.1",
"elasticio-node": "^0.0.8"
}
Building components in Node.js
If you plan to build a component for elastic.io platform in Node.js then you can visit our dedicated documentation page which describes how to build a component in node.js.
Before you start
Before you can deploy any code into our system you must be a registered elastic.io platform user. Please see our home page at http://www.elastic.io to learn how.
Any attempt to deploy a code into our platform without a registration would fail.
After the registration and opening of the account you must upload your SSH Key into our platform.
If you fail to upload you SSH Key you will get permission denied error during the deployment.
Getting Started
After registration and uploading of your SSH Key you can proceed to deploy it into our system. At this stage we suggest you to:
- Create a team to work on your new component. This is not required but will be automatically created using random naming by our system so we suggest you name your team accordingly.
- Create a repository inside the team to deploy your new component code.
Examples of Node.js components
Here is a list of components build on Node.js:
- petstore-component-nodejs to build your first component
- code-component to run pieces of synchronous JavaScript inside your integration flow,
- webhook-component to send and receive WebHooks on elastic.io platform,
- csv-component to work with CSV files in your integration flow,
- sugarcrm-component to use Sugar CRM in your integrations
Flow control
When working in the multi-tenant integration environment it's important to obey the API and consumption limits imposed by the platform. This is not only a condition for you integrations to run on the platform (and not begin suspended), but also a good integration practice to sustainably and efficiently use resources.
Imagine a system where one party (published) publishes to the queue and one or more consumers consume from the queue. If publishers are writing to the queue faster than consumers read data from the queue, queue will earlier or later be overfilled. Once one queue of your integration flow will grow to a particular limit, the complete integration flow will be suspended and author will be informed about it. Flow control is a build-in mechanism in the SDK that will help you to prevent the overflow to happen.
There are two types of flow control:
- Static flow control - the hard limit of the events (e.g. messages published to the queue) that can be generated by component. This limit is dictated by your pricing plan and will limit protect the platform from extensive load.
- Dynamic flow control - the limit that is imposed based on the state of individual queue, more messages are in the queue, slower publisher could write to it.
Let's take a look at the simple example:
'use strict';
exports.process = process;
async function process(msg, cfg, snapshot) {
for (let i = 0; i < 100000; i++) {
console.log('Sending message %s', i);
await this.emit('data', {
body: {
counter: i,
hi: 'there'
}
});
console.log('Message %s was sent', i);
}
}
This simple component, once started on the platform will generate 100k messages. Without flow control this example will quickly bring the integration queue to the limits and integration flow will be suspended. With flow control the publishing rate of the messages will be slowed down so both publisher and consumers will operate in balance.
How to configure it
There is a set of environment variables that are responsible for the configuration of the static flow control (dynamic flow control is implemented in the message-oriented middleware of the platform hence can't be configured on the component level)
- ELASTICIO_DATA_RATE_LIMIT - a number of maximum
data
messages per second that could be emitted by the component
Sailor hooks
Init hook
/**
* cfg - This is the same config as the one passed to "processMessage" method of the trigger or action
*/
exports.init = function(cfg) {
//do stuff
return Promise.resolve();
}
Startup hook
/**
* cfg - This is the same config as the one passed to "processMessage" method of the trigger or action
*/
exports.startup = function(cfg) {
//do stuff
const data = {
message: 'Hello from STARTUP_HOOK'
};
return Promise.resolve(data);
}
Only on the first trigger
Called without
this
May return promise
May return value
May throw - not recommended
May return a promise that will fail
Startup logs can be found in the tab of the component on execution page
TBD - Separate them under different tab in UI
TBD - Where to see restart errors?overwritten
Startup state data - either return value or the result of the promise
- OK
- Results will be stored as the startup state, previous will be overwritten with warning
- After that init hook will be run, etc
- NOK
- Sailor will exit the process
- Platform will restart the component immediately
- If init wont' happen it will be removed after 5 minutes (see restart policy)
- In the next scheduling interval initialisation will repeat
Shutdown hook
/**
* cfg - This is the same config as the one passed to "processMessage" method of the trigger or action
* startData - result from the startup
*/
exports.shutdown = function(cfg, startData) {
//corresponding to the startup example above, startData is { message: 'Hello from STARTUP_HOOK' }
//do stuff
return Promise.resolve();
}
- Only on the first trigger
- One stop is pressed
- If task is running then containers are shutdown
- If task is sleeping then do nothing
- Start new trigger container
- Trigger starts without
this
context - it's not possible to log errors or send new data - Should either return value (ignored) or promise (resolved).
- Startup data is removed after shutdown hook
- Call the shutdown hook, parameters that are passed is from the startup results or
{}
if nothing was returned - Errors are ignored
- If shutdown hook won't complete within 60 seconds then container will be killed
- As soon as user pressed stop, task is marked as inactive and 'webhooks gateway' will start responding with the error (Task either does not exist or is inactive) to possible data
TBD - log for shutdown hooks?