gliner
v0.0.18
Published
This is a GLiNER inference engine
Downloads
88
Readme
👑 GLiNER.js: Generalist and Lightweight Named Entity Recognition for JavaScript
GLiNER.js is a TypeScript-based inference engine for running GLiNER (Generalist and Lightweight Named Entity Recognition) models. GLiNER can identify any entity type using a bidirectional transformer encoder, offering a practical alternative to traditional NER models and large language models.
🌟 Key Features
- Flexible entity recognition without predefined categories
- Lightweight and fast inference
- Easy integration with web applications
- TypeScript support for better developer experience
🚀 Getting Started
Installation
npm install gliner
Basic Usage
const gliner = new Gliner({
tokenizerPath: "onnx-community/gliner_small-v2",
onnxSettings: {
modelPath: "public/model.onnx",
executionContext: "web",
executionProvider: "webgpu",
},
maxWidth: 12,
});
await gliner.initialize();
const input_text = "Your input text here";
const texts = [input_text];
const entities = ["city", "country", "person"];
const threshold = 0.1;
const decoded = await gliner.inference({ texts, entities, threshold });
console.log(decoded);
Advanced Usage
ONNX settings API
- modelPath: can be either a URL to a local model as in the basic example, or it can also be the Model itself as an array of binary data.
- executionProvider: these are the same providers that ONNX web supports, currently we allow
webgpu
(recommended),cpu
,wasm
,webgl
but more can be added - wasmPaths: Path to the wasm binaries, this can be either a URL to the binaries like a CDN url, or a local path to a folder with the binaries.
- multiThread: wether to multithread at all, only relevent for wasm and cpu exeuction providers.
- multiThread: When choosing the wasm or cpu provider, multiThread will allow you to specify the number of cores you want to use.
- fetchBinary: will prefetch the binary from the default or provided wasm paths
🛠 Setup & Model Preparation
To use GLiNER models in a web environment, you need an ONNX format model. You can:
- Search for pre-converted models on HuggingFace
- Convert a model yourself using the official Python script
Converting to ONNX Format
Use the convert_to_onnx.py
script with the following arguments:
model_path
: Location of the GLiNER modelsave_path
: Where to save the ONNX filequantize
: Set to True for IntU8 quantization (optional)
Example:
python convert_to_onnx.py --model_path /path/to/your/model --save_path /path/to/save/onnx --quantize True
🌟 Use Cases
GLiNER.js offers versatile entity recognition capabilities across various domains:
- Enhanced Search Query Understanding
- Real-time PII Detection
- Intelligent Document Parsing
- Content Summarization and Insight Extraction
- Automated Content Tagging and Categorization ...
🔧 Areas for Improvement
- [ ] Further optimize inference speed
- [ ] Add support for token-based GLiNER architecture
- [ ] Implement bi-encoder GLiNER architecture for better scalability
- [ ] Enable model training capabilities
- [ ] Provide more usage examples
Creating a PR
- for any changes, remember to run
pnpm changeset
, otherwise there will not be a version bump and the PR Github Action will fail.
🙏 Acknowledgements
📞 Support
For questions and support, please join our Discord community or open an issue on GitHub.