npm package discovery and stats viewer.

Discover Tips

  • General search

    [free text search, go nuts!]

  • Package details

    pkg:[package-name]

  • User packages

    @[username]

Sponsor

Optimize Toolset

I’ve always been into building performant and accessible sites, but lately I’ve been taking it extremely seriously. So much so that I’ve been building a tool to help me optimize and monitor the sites that I build to make sure that I’m making an attempt to offer the best experience to those who visit them. If you’re into performant, accessible and SEO friendly sites, you might like it too! You can check it out at Optimize Toolset.

About

Hi, 👋, I’m Ryan Hefner  and I built this site for me, and you! The goal of this site was to provide an easy way for me to check the stats on my npm packages, both for prioritizing issues and updates, and to give me a little kick in the pants to keep up on stuff.

As I was building it, I realized that I was actually using the tool to build the tool, and figured I might as well put this out there and hopefully others will find it to be a fast and useful way to search and browse npm packages as I have.

If you’re interested in other things I’m working on, follow me on Twitter or check out the open source projects I’ve been publishing on GitHub.

I am also working on a Twitter bot for this site to tweet the most popular, newest, random packages from npm. Please follow that account now and it will start sending out packages soon–ish.

Open Software & Tools

This site wouldn’t be possible without the immense generosity and tireless efforts from the people who make contributions to the world and share their work via open source initiatives. Thank you 🙏

© 2024 – Pkg Stats / Ryan Hefner

hugbot

v1.5.4

Published

Chatbot maker for HuggingFace Inference API and other AI API providers and backends.

Downloads

11

Readme

HugBot 🤖

Chatbot maker for HuggingFace 🤗 Inference API and other AI API providers and backends.

Features

  • ✨ Free and Open Sauce.
  • 💬 Conversation memory buffer with token counting and truncation to fit the chatbot context window.
  • 🛠️ Fully customizable including system prompt, LLM hyperparameters, prompt tags etc.
  • 🧩 Proper LLM prompt templating using tags and examples provided in model docs and tokenizer configs.
  • 🤖 Preconfigured chatbot models.
  • ⭐️ Works in browsers, Node and Bun.
  • ⚡️ Event driven API with message queue for non-blocking interactions with your bots.
  • 🔥 Modular and extensibe architecture.
  • 📘 All the Typescript typings.
  • 🔮 Integration with different AI api providers other than HuggingFace free API coming soon, maybe.

Preconfigured chatbot models

  • Zephyr: huggingface.co/models/HuggingFaceH4/zephyr-7b-beta
  • Hermes: huggingface.co/models/NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO
  • Mixtral: huggingface.co/models/mistralai/Mixtral-8x7B-Instruct-v0.1
  • Mistral: huggingface.co/models/mistralai/Mistral-7B-Instruct-v0.2
  • StarChat: huggingface.co/models/HuggingFaceH4/starchat2-15b-v0.1

Install

npm i hugbot

Usage

// Import botStorage container:
import { botStorage } from "hugbot"

// Get your bot from bot storage by name:
const zephyr = botStorage.get("Zephyr");

// Use respondTo method to interact with the bot.
// respondTo returns a Promise with generated AI response string:
const response = await zephyr.respondTo("Hi!");
console.log(response);

// Optionally use api token:
// You can get HuggingFace Inference API token here for free: https://huggingface.co/settings/tokens
const response = await zephyr.respondTo("Hi!", apiToken);

// Or use event driven API:
// Add your api token to the bot instance, it will be encrypted and 
// obfuscated in memory during runtime:
await zephyr.apiToken(apiToken);

// Add event listeners for chatbot responses:
zephyr.onResponse((res) => console.log(res));

// Push messages to the queue.
// Messages will be processed one after onother by the bot as 
// they are added to the queue, provided callbacks will be called as the bot responds:
zephyr.pushMessage("hi");
zephyr.pushMessage("?");

// Use setParams method to adjust bot params during runtime:
zephyr.setParams({
  contextWindow: 2000,
  temperature: 0.7,
  systemPrompt: "You are a helpful AI assistant."
});

Use HugBot in Node.js

const hugbot = require('hugbot');

const botStorage = hugbot.botStorage;

const zephyr = botStorage.get("Zephy");

async function chat() {
  const query = prompt(">");
  const res = await zephyr.respondTo(query, "your api token here");
  console.log("\n", res, "\n");
  chat();
}

chat();

Params

Params available for setParams method:

type HugBotParams = {
  systemPrompt: string; // An instruction to AI added to the beginig of the prompt string 
  // example: "You are a helpful AI assistant."
  responseAffirmation: string // Prepended to bot replies, can be used to coerce the bot into following 
  // any instructions, example: "Sure!", "Here you go:". Default is empty string.
  userInstruction: string // Added after user query, can be used for RAG and additional instructions. 
  // Default is empty string.
  // These are only added to last conversation entries.
  contextWindow: number // Chatbot conversation memory size in tokens (around 1.5 tokens per word). 
  // Used to manage limited LLM context window. When memory buffer overflows it's truncated 
  // by removing oldest conversation entries.
  topK: number | undefined;  // Top-K sampling. The range of candidate tokens to select from for the next prediction.
  topP: number | undefined;  // Sampling based on probability threshold.
  temperature: number;  // Parameter impacting the randomness of predictions by scaling the probabilities of alternatives.
  repetitionPenalty: number | undefined;  // Penalizing repeated phrases or sentences by lowering their likelihood.
  maxNewTokens: number | undefined;  // limits the amount of newly generated text per response.
  maxTime: number | undefined;  // Maximum time allowed for generation.
  doSample: boolean;  // Choosing between deterministic greedy decoding (false) and stochastic sampling (true).
}

Example, basic default settings:

bot.setParams({
  systemPrompt: "You are a helpful AI assistant.",
  userInstruction: "",
  responseAffirmation: "",
  contextWindow: 2048,
  maxNewTokens: 500,
  repetitionPenalty: 1.1,
  doSample: true,
  temperature: 0.7,
  topK: 50,
  topP: 0.95,
  maxTime: 30
});

Build your own HugBot

// Import required components and the builder function:
import {  
  BuildHugBot, generateTextResponse, HuggingFaceTextGenClient,
  ShortTermMemory, PromptConstructor, IObuffer, SecretsHider,
  RateLimiter
} from "hugbot"

// Use the builder function to construct and configure the bot.
// All components are optional.
const zephyr = BuildHugBot("Zephyr").fromComponents({
  AIClient: new HuggingFaceTextGenClient({
    languageModel: "HuggingFaceH4/zephyr-7b-beta", // link to the language model from HF
    // part of the full api link
    // full api link looks like this: "https://api-inference.huggingface.co/models/HuggingFaceH4/zephyr-7b-beta"
    maxNewTokens: 256,
    doSample: true,
    temperature: 0.7,
    topK: 50,
    topP: 0.95,
  }),
  shortTermMemory: new ShortTermMemory({
    contextWindow: 2048,
  }),
  promptConstructor: new PromptConstructor({
    bos: "",
    system: "<|system|>\n",
    user: "<|user|>\n",
    bot: "<|assistant|>\n",
    closing: "</s>\n",
  }),
  respondTo: generateTextResponse,
  IObuffer: new IObuffer(),
  secretsHider: SecretsHider(),
  rateLimiter: new RateLimiter(1000)
}).build();

zephyr.respondTo("Hi!").then((response) => console.log(response));

Components

All available HugBot components so far:

 type HugBotEntity = {
  id: string;
  AIClient?: AIClient;
  shortTermMemory?: ShortTermMemory;
  promptConstructor?: PromptConstructor;
  respondTo?: GenerateResponse;
  IObuffer?: IObuffer;
  secretsHider?: SecretsHider;
  rateLimiter?: RateLimiter;
}

export {
  generateTextResponse, BuildHugBot, HugBotProxy,  BotStorage, mistralTokenizer,
  SecretsHider, HuggingFaceTextGenClient, AIClientMock, ShortTermMemory,
  PromptConstructor, MistralPromptConstructor, IObuffer, FCFSqueue, RateLimiter,
}

BuildHugBot

Chatbot builder function for constructing and configuring bots, takes component instances and returns proxied HugBot entity.

const BuildHugBot: (id: string) => {
    fromComponents: (components: Partial<HugBotComponents>) => {
        build: () => HugBot;
    };
}

===>

interface HugBot {
  id: string;
  apiToken: (key: string | null) => Promise<Res>;
  respondTo: (userInput: string, apiToken?: string) => Promise<string>;
  onResponse: (...cb: Array<(res: string) => void>) => void;
  pushMessage: (msg: string) => void;
  setParams: (params: Partial<HugBotParams>) => void;
}

generateTextResponse

Main method for interacting with HugBot. Runs all bot components to produce ai response. Takes HugBot instance, user prompt and optional api token and returns string promise.

type GenerateResponse = (HugBot: HugBotEntity, userInput: string, apiToken?: string) => Promise<string>;

ShortTermMemory

Conversation memory buffer with tokenizer to count prompt tokens and truncate memory entries to fit the prompt inside limited LLM token window.

Uses push() method to add memory entries:

type MemoryEntry = {
  role: "user" | "ai";
  input: string;
}

Dumps current memory state with dump() method to be passed onto prompt constructor:

type MemoryDump = {
  conversation: MemoryEntry[];
  systemPrompt: string;
  responseAffirmation: string;
  userInstruction: string;
}

You can set the following memory state properties:

  • tokenizer - Tokenzer function used for token counting, can be set in constructor, takes string returns num, Mistral tokenizer is used by default;
  • contextWindow - Memory buffer length in tokens;
  • systemPrompt - "You are a helpful AI assistant.";
  • responseAffirmation - Prepended to bot replies, can be used to coerce the bot into following any instructions, example: "Sure!", "Here you go:"
  • userInstruction - added after user query, can be used for RAG and additional instructions.

Short term memory interface:

interface ShortTermMemory {
  push: (entry: MemoryEntry) => void;
  get dump(): MemoryDump;
}

PromptConstructor

Used for prompt templating, takes an object describing ai prompt tags and current short term memory state dump:

tags: {
  bos: "",
  system: "<|im_start|>system\n",
  user: "<|im_start|>user\n",
  bot: "<|im_start|>assistant\n",
  closing: "<|im_end|>\n",
}

Returns prompt template string =>

 "<|im_start|>system
 You are a helpful AI assistant.<|im_end|>
 <|im_start|>user
 hi<|im_end|>
 <|im_start|>assistant
 Hello, how can I help you?<|im_end>
 <|im_start|>user
 Tell me something interesting<|im_end|>
 <|im_start>assistant"

Interface:

interface PromptConstructor {
  getPromptTemplate: (memoryDump: MemoryDump) => string;
}

MistralPromptConstructor

Dedicated Mistral prompt formatter because Mistrals have weird prompt format.

Returns prompt string that looks like this:

"<s>[INST] {system_prompt}\n Hello, how are you? [/INST] I'm doing great. How can I help you today?</s> [INST] I'd like to show off how chat templating works! [/INST]"

HuggingFaceTextGenClient

http client for HuggingFace Inference api. You can find detailed description for all HuggingFace API params here: https://huggingface.co/docs/api-inference/detailed_parameters

Available props:

type HuggingFaceTextGenParams = {
  endPoint: string; // link to HuggingFace api end point: "https://api-inference.huggingface.co/models/"
  languageModel: string; // link to the language model, assigned in constructor.
  // example: "HuggingFaceH4/zephyr-7b-beta"
  // full api link example used to fetch ai responses: 
  // "https://api-inference.huggingface.co/models/HuggingFaceH4/zephyr-7b-beta"
  // LLM hyperparameters:
  topK: number | undefined;
  topP: number | undefined;
  temperature: number;
  repetitionPenalty: number | undefined;
  maxNewTokens: number | undefined;
  maxTime: number | undefined;
  returnFullText: boolean;
  numReturnSequences: number;
  doSample: boolean;
  truncate: number | undefined;
  // Server options:
  waitForModel: boolean;
  useCache: boolean;
}

HuggingFace client exposes this interface to integrate with HugBot:

interface AIClient {
  sendRequest: (prompt: string, apiToken?: string) => Promise<string>;
}

AIClientMock

Fake AI client for testing purposes, returns random responses with random delays.

IObuffer

This is used to implement event driven API with message queue. The message queue is processed asynchronously by the bot. You can assign multiple event listeners to handle bot responses using onResponse method. Use pushMessage to add messages to the queue. IObuffer uses rate limiter, you can set minimum delay between respondTo calls with rateLimit property in constructor.

interface IObuffer {
  pushMessage: (msg: string) => void;
  onResponse: (...cb: Array<(res: string) => void>) => void;
  removeEventListener: (name: string | "all") => void;
  set setBot(bot: HugBotEntity);
}

HugBotProxy

Proxy layer for input validation, handling access and setters. Wraps HugBot entity instance before returning it from the builder function with .build() method.

const HugBotProxy: (bot: HugBotEntity) => HugBotEntity

SecretsHider

Used to encrypt and obfuscate provided api key in memory during runtime.

interface SecretsHider {
  set: (secret: string) => Promise<Res>;
  get: () => Promise<string | null | Res>;
  destroy: () => void;
}

RateLimiter

Built in rate limiter module to prevent API abuse. Throws if spam is detected. Constructor takes minimum delay betwen requests in milliseconds.

interface RateLimiter {
  check: () => void | never;
}

BotStorage

Bot storage container. Bot names mapped to their corresponding builder functions. The bot builder function is called and the bot is instantiated when it's retrieved from storage using the get method.

type BotsInStorage = "StarChat" | "Zephyr" | "Hermes" | "Mixtral" | "Mistral";

const BotStorage: () => {
    put: (id: string, fn: () => HugBot) => {
        put: ...;
        get: (id: BotsInStorage) => HugBot;
    };
    get: (id: BotsInStorage) => HugBot;
}

License

This project is licensed under the MIT License.