npm package discovery and stats viewer.

Discover Tips

  • General search

    [free text search, go nuts!]

  • Package details

    pkg:[package-name]

  • User packages

    @[username]

Sponsor

Optimize Toolset

I’ve always been into building performant and accessible sites, but lately I’ve been taking it extremely seriously. So much so that I’ve been building a tool to help me optimize and monitor the sites that I build to make sure that I’m making an attempt to offer the best experience to those who visit them. If you’re into performant, accessible and SEO friendly sites, you might like it too! You can check it out at Optimize Toolset.

About

Hi, 👋, I’m Ryan Hefner  and I built this site for me, and you! The goal of this site was to provide an easy way for me to check the stats on my npm packages, both for prioritizing issues and updates, and to give me a little kick in the pants to keep up on stuff.

As I was building it, I realized that I was actually using the tool to build the tool, and figured I might as well put this out there and hopefully others will find it to be a fast and useful way to search and browse npm packages as I have.

If you’re interested in other things I’m working on, follow me on Twitter or check out the open source projects I’ve been publishing on GitHub.

I am also working on a Twitter bot for this site to tweet the most popular, newest, random packages from npm. Please follow that account now and it will start sending out packages soon–ish.

Open Software & Tools

This site wouldn’t be possible without the immense generosity and tireless efforts from the people who make contributions to the world and share their work via open source initiatives. Thank you 🙏

© 2024 – Pkg Stats / Ryan Hefner

suffix-thumb

v5.0.2

Published

learn transformations between two sets of words

Downloads

285,550

Readme

discover the minimal rules for mapping two sets of words to one another, according to changes in their suffix.

It was built for learning rules about verb conjugations, but in a way, it is just a generic compression algorithm.

The assumption is that a word's suffix is the most-often changed part of a word.

preview

Learn → Convert

import { learn, convert } from 'suffix-thumb'

let pairs = [
  ['walk', 'walked'],
  ['talk', 'talked'],
  ['go', 'went'],
]
let model = learn(pairs)
/* {
  rules: { k: [ [ 'alk', 'alked' ] ] },
  exceptions: { go: 'went' },
}*/

let out = convert('walk', model)
// 'walked'

you can pass-in options:

let opts={
  threshold:80, //how sloppy our initial rules can be
  min:0, //rule must satisfy # of pairs
  reverse:true, //compute backward transformation, too
}
let model = learn(pairs, opts)

Reverse

the model also works transforming the words the other way:

import { learn, reverse, convert } from 'suffix-thumb'

let pairs = [
  ['walk', 'walked'],
  ['talk', 'talked'],
  ['go', 'went'],
]
let model = learn(pairs)
let rev = reverse(model)
let out = convert('walked', rev)
// 'walk'

by default, the model ensures all two-way transformation - if you only require 1-way, you can do:

learn(pairs, {reverse: false})

you can expect the model to be 5% smaller or so - not much.

Compress

by default, the model is small, but remains human-readable (and human-editable). We can compress it further, turning it into a snowball inscrutible characters:

import { learn, compress, uncompress, convert } from 'suffix-thumb'

let pairs = [
  ['walk', 'walked'],
  ['talk', 'talked'],
  ['go', 'went'],
]
let model = learn(pairs)
// shrink it
model = compress(shrink)
// {rules:'LSKs3H2-LNL.S3DH'}
// pop it back
model = uncompress(model)
let out = convert('walk', model)
// 'walked'

The models must be uncompressed before they are used, or reversed.

Validation

sometimes you can accidentally send in an impossible set of transformations. This library quietly ignores duplicates, by default. You can use {verbose:true} to log warnings about this, or validate your input manually:

import { validate } from 'suffix-thumb'
let pairs = [
  ['left', 'right'],
  ['left', 'right-two'],
  ['ok', 'right'],
]
pairs = validate(pairs) //remove dupes (on both sides)

If you are just doing one-way transformation, and not reverse, you may want to allow duplicates on the right side:

let pairs = [
  ['left', 'right'],
  ['ok', 'right'],
]
let model = learn(pairs, {reverse: false})
let out = convert('ok', model)
// 'right'

How it works

For each word-pair, it generates all n-suffixes of the left-side, and n-suffixes of the right-side.

any good correlations between the two suffix pairs begins to pop out. Exceptions to these rules are remembered. It then exhaustively reduces any redundancies in these rules.

There are some compromises, magic-numbers, and opinionated decisions - in-order to allow productive, but imperfect rules.

  • The library is meant optimize for file-size of the model
  • compression is slow, uncompression is fast
  • it should always return a perfect result

The library drops case-information - and numbers and some characters1 will not compress properly.

There may be wordlists with few helpful patterns. Conjugation datasets in English and French tend to get ~85% filesize compression.

See also

  • efrt - trie-based prefix compression for JSON

MIT