arpk
v0.1.1
Published
Ollama as your translator, with DeepLX-compatible API.
Downloads
399
Readme
ARPK
Ollama as your translator, with DeepLX-compatible API.
Usage
Prepare
You need to run Ollama and download the model you want to use.
https://ollama.com/
ollama serve
ollama pull llama3.1
Run
npm
You can run ARPK directly without installation.
ARPK_MODEL="llama3.1" npx arpk
container
TODO
API
/translate
, /api/v1/translate
and /api/v2/translate
are connected to the same translate endpoint.
await fetch('http://127.0.0.1:1188/translate', {
body: JSON.stringify({
source_lang: 'JA',
target_lang: 'ZH',
text: '雨の季節が過ぎ 澄み渡る空を 眺めて独り想フ'
}),
method: 'POST'
}).then(res => res.json())
// {
// alternates: [],
// code: 200,
// data: '雨季过后,晴朗的天空下我独自遐思。',
// id: 1519129853500,
// method: 'ARPK',
// source_lang: 'JA',
// target_lang: 'ZH'
// }
Environments
Currently only Bearer Auth is supported when using
ARPK_TOKEN
, not URL Params.
| Environment | Default | Description | |--------------------|---------------------------------------------------------------|------------------------------------| | ARPK_PORT | 1188 | The port the server will listen on | | ARPK_MODEL | llama3.1 | Model to be used by the ARPK | | ARPK_TOKEN | null | Access token to protect your API | | ARPK_OLLAMA_HOST | http://127.0.0.1:11434 | The Ollama host address | | ARPK_SYSTEM_PROMPT | https://github.com/moeru-ai/arpk/blob/main/src/lib/prompts.ts | System prompt |