epha-robot
v0.2.42
Published
Fetching, cleaning, transforming of pharmaceutical data from public resources
Downloads
60
Maintainers
Readme
epha-robot
robot is a tool (entirely written in node.js) for fetching, purifying and transforming pharmaceutical (csv, xlsx, xml, zip) into machine readable (JSON, csv) data.
robot uses public resources like swissmedic - Product information and is meant to be as a starting point for studies, thesis and further processing and purifying of the data.
Table of contents
Benefits
- reliable and smart fetching of pharmaceutical data
- auto-transformation into JSON-files: for example from xlsx-files
- supports the following data/sources:
Jobs
atc
This job generates a map of Anatomical Therapeutic Chemical Classification System-data.
Start:
- go to robot location and type
npm start
npm start
> [email protected] start /Your/robot/location/
> node ./bin/cli.js
EMIL: I'm ready, if you are? Type help for help.
- after prompt is ready type atc
> atc
EMIL: Added 'atc' to the queue (1 jobs)!
EMIL: You can run queue with 'go'
- then type go to start queued job
> go
- ... some logging ...
[email protected] | TIME | ATC Completed in { duration: '15306ms' }
- ... done!
Downloads:
- source: WIdO - Wissenschaftliches Institut der AOK
{PROCESS_ROOT}/data/auto/atcatc.zip
(> 4.5MB) containing atc.xlsx
Releases:
- drive:
{PROCESS_ROOT}/data/release/atc
atc.csv
atc.json
atc.min.json
atc.json - Sample:
//..
"A01AA51": {
"name": "Natriumfluorid, Kombinationen"
},
"A01AB": {
"name": "Antiinfektiva und Antiseptika zur oralen Lokalbehandlung"
},
"A01AB02": {
"name": "Wasserstoffperoxid",
"ddd": "60 mg O"
},
//..
bag
Gets a collection of pharmaceutical products containing purchase and selling price. There is also a history keeping track of all products (incl. de-registered products). Besides that the job also provides bi-temporal data for purchase and selling prices.
Start:
- go to robot location and type
npm start
npm start
> [email protected] start /Your/robot/location/
> node ./bin/cli.js
EMIL: I'm ready, if you are? Type help for help.
- after prompt is ready type bag
> bag
EMIL: Added 'bag' to the queue (1 jobs)!
EMIL: You can run queue with 'go'
- then type go to start queued job
> go
- ... some logging ...
[email protected] | TIME | BAG Completed in { duration: '28844ms' }
- ... done!
Downloads:
- source: BAG - Bundesamt für Gesundheit (CH)
- drive:
{PROCESS_ROOT}/data/auto/bag/XMLPublications.zip
(~ 5MB) contains:bag.xls
,bag.xml
,it.xml
Releases:
- drive:
{PROCESS_ROOT}/data/release/bag/
bag.json
bag.min.json
bag.history.json
bag.history.min.json
bag.price-history.json
bag.price-history.min.json
it.json
it.min.json
bag.json - Sample:
// ...
{
"name": "3TC",
"atc": "J05AF05",
"description": "Filmtabl 150 mg",
"orgGenCode": "O",
"flagSB20": "N",
"vatInEXF": "N",
"substances": [
{
"name": "Lamivudinum",
"quantity": "150",
"quantityUnit": "mg"
}
],
"packung": "60 Stk",
"flagNarcosis": "N",
"bagDossier": "16577",
"gtin": "7680536620137",
"exFactoryPreis": "164.55",
"exFactoryPreisValid": "01.10.2011",
"publikumsPreis": "205.30",
"publikumsPreisValid": "01.10.2011",
"validFrom": "15.03.1996"
},
//...
bag-history(-job)
In bag.history.json
the job keeps automatically track of de-registered products and price changes. This file will be automatically created after the first run (at this moment contents will be equal to bag.history.json
). Deleting this file is the same as restarting the history. Probably it is necessary - especially bevor un-installing/removing robot - to backup this file from time to time.
bag.history.json - Sample:
//...
"publikumsPreisHistory": [
// history-entity
{
"dateTime": "08.06.2015 17:09", // time of change
"publikumsPreis": [
"205.30", //before
"300.00" //after
],
"publikumsPreisValid": [
"01.10.2011", //before
"08.06.2015" //after
]
}
// ..
]
//...
bag-price-history
robot records in bag.price-history.json
product price changes (purchase and selling price). Each run of the job will update this file if a change was detected.
Products are identified by their GTIN.
Usually prices rarely change. So dates at validFrom
and validTo
are on day basis. Date is formatted like in bag.json
: DD.MM.YYYY
. Please note: validFrom
is including while validTo
is excluding.
There are two types of prices:
exFactory
: purchase pricepublikum
: selling price
and two sub-types:
valid
: time for a price valid in the real worldtransaction
: time for a price detected by robot
valid
and transaction
are collections and latest price may be found at index 0
. validTo
is null
or rather Infinite
for most recent price as this information is not available.
bag.price-history.json - Sample:
{
"7680536620137": [
{
"exFactoryPreis": "196.35",
"publikumsPreis": "214.99",
"validFrom": "18.06.2015",
"validTo": null,
"transactionFrom": "18.06.2015", // recorded by robot
"transactionTo": null
},
{
"exFactoryPreis": "176.45",
"publikumsPreis": "209.99",
"validFrom": "11.01.2011", // parsed from data
"validTo": "17.06.2015",
"transactionFrom": "01.01.2015", // recorded by robot
"transactionTo": "17.06.2015"
}
],
}
bag-logs
Additionally to the history-file, logs for new, changed and de-registered products will be written:
- drive:
{PROCESS_ROOT}/logs/bag/
bag.changes.log
bag.new.log
bag.de-registered.log
It could be very handy to use tail -f
on this logs.
kompendium:
The kompendium-job fetches a huge catalog of pharmaceutical product information and is also quite time and resource consuming. The downloaded file itself has around 190MB (> 800MB unzipped). The job will also build a huge amount of .htm-files (~25000) containing product specific and patient related information in German, French and Italian (if available).
Start:
- go to robot location and type
npm start
npm start
> [email protected] start /Your/robot/location/
> node ./bin/cli.js
EMIL: I'm ready, if you are? Type help for help.
- after prompt is ready type kompendium
> kompendium
EMIL: Added 'kompendium' to the queue (1 jobs)!
EMIL: You can run queue with 'go'
- then type go to start queued job
> go
- ... some logging ...
[email protected] | TIME | Kompendium Completed in { duration: '299261ms' }
- ... done!
Downloads
- source: Swissmedic - Swiss Agency for Therapeutic Products
- drive:
{PROCESS_ROOT}/data/auto/kompendium/kompendium.zip
(190MB) - containing
kompendium.xml
(~850MB)
Releases
- drive:
{PROCESS_ROOT}/data/release/kompendium
kompendium.json
kompendium.min.json
catalog.json
- German FI/PI:
{PROCESS_ROOT}/data/release/kompendium/de/
fi/{REGISTRATION_NUMMBER}.htm
pi/{REGISTRATION_NUMMBER}.htm
- French FI/PI:
{PROCESS_ROOT}/data/release/kompendium/fr/
fi/{REGISTRATION_NUMMBER}.htm
pi/{REGISTRATION_NUMMBER}.htm
- Italian FI/PI:
{PROCESS_ROOT}/data/release/kompendium/it/
fi/{REGISTRATION_NUMMBER}.htm
pi/{REGISTRATION_NUMMBER}.htm
kompendium.json - Sample:
{
"documents": [
// ...
{
"zulassung": "10167",
"lang": "de fr it",
"type": "fi pi",
"produkt": "Emser Salz®",
"substanz": "Emser Salz",
"hersteller": "Sidroga AG",
"atc": "RO2AX",
"files": [
//language/type/{REGISTRATION_NUMMBER}.htm
"de/fi/10167.htm",
"fr/fi/10167.htm",
"de/pi/10167.htm",
"fr/pi/10167.htm",
"it/pi/10167.htm"
]
}
// ...
]
}
swissmedic:
This job fetches data about human and veterinary medicines. It also creates a history-file and triggers the atc-Job if required.
atc/CH
When there is no atc-Release available it auto-runs the atc-Job as it is a dependency for atcCH. Please note that if there is an atc-Release available it will use it. This release could be potentially out-of-date. So it is up to the user to run atc-Job if necessary.
swissmedicHistory(-job)
There will be also a swissmedic.history.json
which keeps track of de-registered products. This file will be automatically created after the first run (at that moment contents will be equal to swissmedic.json
). Deleting this file is the same as restarting the history. De-registered products will be flagged with { "deregistered": "DD.MM.YYYY" }
. Please note: Before re-installing robot it is advisible to backup this file.
swissmedic-logs
Like bag there will be logs for new, changed and de-registered products:
- drive:
{PROCESS_ROOT}/logs/swissmedic/
swissmedic.changes.log
swissmedic.new.log
swissmedic.de-registered.log
Think of tail -f
, it might be useful.
Start:
- go to robot location and type
npm start
npm start
> [email protected] start /Your/robot/location/
> node ./bin/cli.js
EMIL: I'm ready, if you are? Type help for help.
- after prompt is ready type swissmedic
> swissmedic
EMIL: Added 'swissmedic' to the queue (1 jobs)!
EMIL: You can run queue with 'go'
- then type go to start queued job
> go
- ... some logging ...
[email protected] | TIME | Swissmedic Completed in { duration: '13369ms' }
- ... done!
Downloads:
source: Swissmedic - Swiss Agency for Therapeutic Products
swissmedic:
- location:
{PROCESS_ROOT}/data/auto/swissmedic/
swissmedic.xlsx
(> 2.5MB)
- location:
atc (as a side effect)
- location:
{PROCESS_ROOT}/data/release/atc
atc.zip
(> 4.5MB), containingatc.xlsx
- location:
Releases:
atc
- location:
{PROCESS_ROOT}/data/release/atc
atc_de-ch.json
atc_de-ch.min.json
- as a side effect
- (
atc.csv
) - (
atc.json
) - (
atc.min.json
)
- (
- location:
swissmedic:
- location:
{PROCESS_ROOT}/data/release/swissmedic/
swissmedic.json
swissmedic.min.json
swissmedic.history.json
swissmedic.history.min.json
- location:
swissmedic.json - Sample
//..
{
"zulassung": "00277",
"sequenz": "1",
"name": "Coeur-Vaisseaux Sérocytol, suppositoire",
"hersteller": "Sérolab, société anonyme",
"itnummer": "08.07.",
"atc": "J06AA",
"heilmittelcode": "Blutprodukte",
"erstzulassung": "26.4.2010",
"zulassungsdatum": "26.4.2010",
"gueltigkeitsdatum": "25.4.2020",
"verpackung": "001",
"packungsgroesse": "3",
"einheit": "Suppositorien",
"abgabekategorie": "B",
"wirkstoffe": "globulina equina (immunisé avec coeur, endothélium vasculaire porcins)",
"zusammensetzung": "globulina equina (immunisé avec coeur, endothélium vasculaire porcins) 8 mg, propylenglycolum, conserv.: E 216, E 218, excipiens pro suppositorio.",
"anwendungsgebiet": "Traitement immunomodulant selon le Dr Thomas\r\n\r\nPossibilités d'emploi voir information professionnelle",
"gtin": "7680002770014"
},
//..
Install robot
Requirements
- node.js >= v0.12.x (Installing information)
- npm > 2.7.x (usually shipped with node.js)
Installation
npm
npm install epha-robot
github
cd path/to/your/WORKSPACE
git clone https://github.com/epha/robot.git
cd robot
npm install
Usage
CLI
npm start
> [email protected] start
> node ./bin/cli.js
EMIL: I'm ready, if you are? Type help for help.
> help
EMIL: You can add jobs to the queue e.g.
EMIL: 'atc' << Codes & DDD
EMIL: 'bag' << Spezialitätenliste
EMIL: 'kompendium' << Swissmedic Kompendium
EMIL: 'swissmedic' << Registered products CH
EMIL: and then run queue with 'go'
EMIL: I'm ready, if you are? Type help for help.
>
npm scripts
robot-service
npm run robot-service
Probably the most common use-case for robot: runs outdated each 30 minutes (default). It is possible to adjust re-run-time by passing DELAY={OTHER_VALUE}
(milliseconds). DELAY
should depend on your internet connection and cpu power.
Example:
DELAY=60000 npm run robot-service
will run outdated every hour.
Will only exit manually or when it crashes. Log level is reduced to warnings and errors.
It could be quite useful running the underlying script (bin/outdated
) with a daemon like forever, pm2 so that it will automatically restart if it crashes (which shouldn't happen).
stdout - sample:
[email protected] | WARN | robot-service 12.06.2015 08:36 - Start Outdated Check
[email protected] | WARN | BAG File on disk is up-to-date
[email protected] | WARN | ATC File on disk is up-to-date
[email protected] | WARN | Swissmedic File on disk is up-to-date
[email protected] | WARN | Kompendium File on disk is up-to-date
[email protected] | WARN | robot-service 12.06.2015 08:36 - Finished Outdated Check
[email protected] | WARN | robot-service 12.06.2015 09:06 - Start Outdated Check
[email protected] | WARN | BAG File on disk is up-to-date
[email protected] | WARN | ATC File on disk is up-to-date
[email protected] | WARN | Swissmedic File on disk is up-to-date
[email protected] | WARN | Kompendium File on disk is up-to-date
[email protected] | WARN | robot-service 12.06.2015 09:06 - Finished Outdated Check
start
npm start
Starts the robot-cli.
all
npm run all
Runs all jobs (atc, bag, kompendium, swissmedic) in parallel. Useful with a broadband internet connection and powerful cpu to get as fast as possible the current state. Exists when done or fails. Will overwrite/updates existing files.
outdated
npm run outdated
Checks sources on changes by header-content-length
. If content-length
diffs to file-size on disk it will trigger appropriate job. Runs jobs in sequence and exits when is done or fails.
Programmatical
var robot = require("epha-robot");
var disk = require("epha-robot").common.disk;
var kompendiumJob = robot.kompendium;
var kompendiumCfg = robot.kompendium.cfg;
kompendiumJob()
.then(function () {
return disk.read.json(kompendiumCfg.process.file);
})
.then(function (data) {
// do something with data
})
.catch(function (err) {
console.error("OH NO!", err.message, err.stack);
});
Development
job-configs
Each job has it's own config file. However there is a convention for configs:
/* any config file*/
// Will resolve paths according to {PROCESS_ROOT}
var config = require("lib/common/config");
module.exports = config("anyJobName", {
"download": {
"url": "..."
"linkParser": /RegExp/i
"zipFiles": [{name: /RegExpForFileInZip/, dest: }]
},
//optional
"manual": {
},
"release": {
"file": "anyJobName.json" // will resolve to {PROCESS_ROOT}/data/anyJobName/release/anyJobName.json
"minFile": "anyJobName.min.json",
"nested": {
"file": nested.json //will also resolve to full path
}
},
//optional
"history": {
"file": "anyJobName.history.json",
"minFile": "anyJobName.history.min.json"
},
//optional
"log": {
"deRegistered": "anyJobName.de-registered.log",
"changes": "anyJobName",
"new": "anyJobName"
}
});
creating a history file
Basically it should be possible to create for each release a history-file by using history
-lib if
- the release is a JSON-collection.
- each collection entry has a key
gtin
that identifies this entry
/* history job for anyJob */
var history = require("lib/history/history");
var cfg = require("jobs/cfg/anyJobCfg");
/**
* Pass a logger if default-logger doesn't fit with your desired log-level, but it is optional
* History returns a Promise.
*/
function anyJobHistory(log) {
// will be called if a change was detected.
// passes references to currently processed history- and newData entry
function onChanged(diff, historyData, newData) {
// do something fancy, a good example mighty be jobs/bagHistory.js
}
// cfg must contain information about where to put history- and log-files
// @see job.configs
return history("anyJob", cfg, onChanged, log);
}
module.exports = anyJobHistory;
working with files
robot ships with lib/common/disk.js
which allows comfortable working with files through a Promise-based-API. This example should give an idea of what it can do:
var path = require("path");
var disk = require("lib/common/disk");
var bagJob = require("jobs/bag");
var bagCfg = require("jobs/cfg/bag.cfg)";
var processBAGData = require("lib/processBAGData");
function workWithFiles() {
return new Promise(function (resolve, reject) {
disk
.fileExists(cfg.download.file)
.then(function (fileExists) {
if (fileExists && path.extname(cfg.download.file) === ".zip") {
// zipFiles is an array with information about files which should be unzipped
return disk
.unzip(cfg.download.zipFiles)
.then(function () {
return disk.read.file(cfg.download.zipFiles[0].name)
})
.then(function (unzippedFileData) {
return processBAGData(unzippedFileData);
})
.then(function (processedData) {
return Promise.all([disk.write.json("myFile.json"), disk.write.jsonMin("myFile.min.json")]);
})
.catch(reject);
}
// else run bag-job first
bagJob().then(workWithFiles);
})
.catch(reject);
});
}
Tests
Unit-Tests
npm test
: Runs the unit-tests oncenpm run watch-test
: Watches project's files and re-runs unit-tests on change- both support growling. check tj/node-growl to enable it on your machine.
Integration-Tests
npm run test-integration
- Run
npm run init-test-integration
which will download fresh data to{ROOT}/data/auto
&{ROOT}/data/release
and copies it to{ROOT}/fixtures
to use it as fixtures - Spins up a real node http-server that serves atc, swissmedic etc. dummy sites and downloads.
- It runs each job against the integration-testing-server and tests the whole flow from html parsing to creating release files.