npm package discovery and stats viewer.

Discover Tips

  • General search

    [free text search, go nuts!]

  • Package details

    pkg:[package-name]

  • User packages

    @[username]

Sponsor

Optimize Toolset

I’ve always been into building performant and accessible sites, but lately I’ve been taking it extremely seriously. So much so that I’ve been building a tool to help me optimize and monitor the sites that I build to make sure that I’m making an attempt to offer the best experience to those who visit them. If you’re into performant, accessible and SEO friendly sites, you might like it too! You can check it out at Optimize Toolset.

About

Hi, 👋, I’m Ryan Hefner  and I built this site for me, and you! The goal of this site was to provide an easy way for me to check the stats on my npm packages, both for prioritizing issues and updates, and to give me a little kick in the pants to keep up on stuff.

As I was building it, I realized that I was actually using the tool to build the tool, and figured I might as well put this out there and hopefully others will find it to be a fast and useful way to search and browse npm packages as I have.

If you’re interested in other things I’m working on, follow me on Twitter or check out the open source projects I’ve been publishing on GitHub.

I am also working on a Twitter bot for this site to tweet the most popular, newest, random packages from npm. Please follow that account now and it will start sending out packages soon–ish.

Open Software & Tools

This site wouldn’t be possible without the immense generosity and tireless efforts from the people who make contributions to the world and share their work via open source initiatives. Thank you 🙏

© 2024 – Pkg Stats / Ryan Hefner

ollama-node

v0.1.28

Published

an easy to use node module for the ollama api

Downloads

669

Readme

I am not really focusing on this anymore. The official library came out a few months ago, and that works a lot better. I am sure more and more of this will break.

Simple JS library to work with Ollama

The simplest JavaScript library for the easiest way to run LLMs.

This is not the official library (we don't have one), but I am one of the maintainers of Ollama.

Get started

import { Ollama } from 'ollama-node';

const ollama = new Ollama();
await ollama.setModel("llama2");

// callback to print each word 
const print = (word: string) => {
  process.stdout.write(word);
}
await ollama.streamingGenerate("why is the sky blue", print);

Create the Ollama Object

const ollama = new Ollama();

Creates the Ollama object. All methods are called from this object.

Optional Parameter

Hostname - defaults to 127.0.0.1

Set Model

await ollama.setModel("llama2";)

Sets the model to use for Generation. Unless you override anything, it will use the template, parameters, and system prompt from the Modelfile.

Working with the Model

Set System Prompt

ollama.setSystemPrompt("You are an AI assistant.");

Sets the system prompt to use with this model. Overrides anything set in the Modelfile.

Set Template

ollama.setTemplate("this is a template")

Add a Parameter

ollama.addParameter("stop", "User:")

Delete a Parameter

ollama.deleteParameter("stop", "User:")

Delete a Parameter by Name

ollama.deleteParameterByName("stop");

Deletes all parameters with that name.

Delete All Parameters

ollama.deleteAllParameters();

Show All Parameters

const params = ollama.showParameters();

Show System Prompt

const sprompt = await ollama.showSystemPrompt()

Useful if you want to update the system prompt based on the existing one.

Show Template

const template = ollama.showTemplate();

Show Model

const model = ollama.showModel();

Shows the current model name

Show Model Info

const info = await ollama.showModelInfo();

Returns parameters, template, system prompt for the current model.

List All the Models Already Pulled

const models = await ollama.listModels();

Lists all local models already downloaded.

Generate (Ask a question and get an answer back)

const output = await ollama.generate("Why is the sky blue?");

This will run the generate command and return the output all at once. The output is an object with the output and the stats.

If you want the streaming version, see below.

Streaming Generate

This is a streaming version of generate, but you don't need to know anything about streaming. Just write a callback function that does what you want to happen.

const printword = (word: string) => {
  process.stdout.write(word);
}

const printline = (line: string)

await ollama.streamingGenerate("why is the sky blue", printword, null, printline)

There are four potential callbacks, all of which are optional, though their positions matter. Use null if you want a later one and not an earlier one.

The Callbacks are:

  • responseOutput: outputs just the token in the response.
  • contextOutput: outputs the context at the end.
  • fullResponseOutput: outputs the full response object.
  • statsOutput: outputs the stats object at the {% endif %}

Other functions I need to document

  • create
  • streamingCreate
  • streamingPull
  • streamingPush
  • copy

This is not in a finished state. It is absolutely a work in progress. I started putting this together and then later saw someone put out another library on npm. bummer. but cool that it's exciting for other folks.

I just want an easy way to consume this in some examples.

I'll flesh it out soon.