npm package discovery and stats viewer.

Discover Tips

  • General search

    [free text search, go nuts!]

  • Package details

    pkg:[package-name]

  • User packages

    @[username]

Sponsor

Optimize Toolset

I’ve always been into building performant and accessible sites, but lately I’ve been taking it extremely seriously. So much so that I’ve been building a tool to help me optimize and monitor the sites that I build to make sure that I’m making an attempt to offer the best experience to those who visit them. If you’re into performant, accessible and SEO friendly sites, you might like it too! You can check it out at Optimize Toolset.

About

Hi, 👋, I’m Ryan Hefner  and I built this site for me, and you! The goal of this site was to provide an easy way for me to check the stats on my npm packages, both for prioritizing issues and updates, and to give me a little kick in the pants to keep up on stuff.

As I was building it, I realized that I was actually using the tool to build the tool, and figured I might as well put this out there and hopefully others will find it to be a fast and useful way to search and browse npm packages as I have.

If you’re interested in other things I’m working on, follow me on Twitter or check out the open source projects I’ve been publishing on GitHub.

I am also working on a Twitter bot for this site to tweet the most popular, newest, random packages from npm. Please follow that account now and it will start sending out packages soon–ish.

Open Software & Tools

This site wouldn’t be possible without the immense generosity and tireless efforts from the people who make contributions to the world and share their work via open source initiatives. Thank you 🙏

© 2024 – Pkg Stats / Ryan Hefner

refa

v0.12.1

Published

A library for finite automata and regular expressions in the context of JS RegExp

Downloads

2,453,037

Readme

Regular Expressions and Finite Automata (refa)

Actions Status npm

A library for regular expressions (RE) and finite automata (FA) in the context of Javascript RegExp.

About

refa is a general library for DFA, NFA, and REs of formal regular languages. It also includes methods to easily convert from JS RegExp to the internal RE AST and vice versa.

Installation

Get refa from NPM:

npm i --save refa

or

yarn add refa

Features

  • Conversions

    • RE AST to NFA and ENFA (assertions are not implemented yet)
    • DFA, NFA, and ENFA can all be converted into each other
    • DFA, NFA, and ENFA to RE AST
  • DFA, NFA, and ENFA operations

    • Construction from other FA, the intersection of two FA, or a finite set of words
    • Print as DOT or Mermaid.
    • Test whether a word is accepted
    • Test whether the accepted language is the empty set/a finite set
    • Accept all prefixes/suffixes of a language
  • DFA specific operations

    • Minimization
    • Complement
    • Structural equality
  • NFA and ENFA specific operations

    • Union and Concatenation with other FA
    • Quantification
    • Reverse
  • AST transformations

    • Simplify and change the AST of a regex
    • Remove assertions
  • JavaScript RegExp

    • RegExp to RE AST and RE AST to RegExp
      • All flags are fully supported
      • Unicode properties
      • Change flags
      • Limited support for simple backreferences

See the API documentation for a complete list of all currently implemented operations.

RE AST format

refa uses its own AST format to represent regular expressions. The RE AST format is language agnostic and relatively simple.

It supports:

  • Concatenation (e.g. ab)
  • Alternation (e.g. a|b)
  • Quantifiers (greedy and lazy) (e.g. a{4,6}, a{2,}?, a?, a*)
  • Assertions (e.g. (?=a), (?<!a))
  • Characters/character sets (represented by interval sets)
  • Unknowns (elements that cannot be represented otherwise. E.g. backreferences)

Some features like atomic groups and capturing groups are not supported (but might be added in the future).

For information on how to parse JS RegExp and convert RE AST to JS RegExp, see the JS namespace.

Universal characters

refa does not use JavaScript strings represent characters or a sequences of characters. Instead it uses integers to represent characters (see the Char type) and arrays of numbers to represent words/strings (see the Word type).

This means that any text encoding can be used.

The Words namespace contains functions to convert JavaScript data into refa-compatible words and characters.

For the sets of characters, the CharSet class is used.

General limitations

This library will never be able to support some modern features of regex engines such as backreferences and recursion because these features, generally, cannot be be represented by a DFA or NFA.

Usage examples

refa is a relatively low-level library. It only provides the basic building blocks. In the following examples, JS RegExps are used a lot so we will define a few useful helper function beforehand.

import { DFA, FiniteAutomaton, JS, NFA } from "refa";

function toNFA(regex: RegExp): NFA {
	const { expression, maxCharacter } = JS.Parser.fromLiteral(regex).parse();
	return NFA.fromRegex(expression, { maxCharacter });
}
function toDFA(regex: RegExp): DFA {
	return DFA.fromFA(toNFA(regex));
}
function toRegExp(fa: FiniteAutomaton): RegExp {
	const literal = JS.toLiteral(fa.toRegex());
	return new RegExp(literal.source, literal.flags);
}
  • toNFA parses the given RegExp and constructs a new NFA from the parsed AST.
  • toDFA constructs a new NFA from the RegExp first and then converts that NFA into a new DFA.
  • toRegex takes an FA (= NFA or DFA) and converts it into a RegExp.

Testing whether a word is accepted

import { Words } from "refa";

const regex = /\w+\d+/;
const nfa = toNFA(regex);

console.log(nfa.test(Words.fromStringToUTF16("abc")));
// => false
console.log(nfa.test(Words.fromStringToUTF16("123")));
// => true
console.log(nfa.test(Words.fromStringToUTF16("abc123")));
// => true
console.log(nfa.test(Words.fromStringToUTF16("123abc")));
// => false

Finding the intersection of two JS RegExps

const regex1 = /a+B+c+/i;
const regex2 = /Ab*C\d?/;

const intersection = NFA.fromIntersection(toNFA(regex1), toNFA(regex2));

console.log(toRegExp(intersection));
// => /Ab+C/

Finding the complement of a JS RegExp

const regex = /a+b*/i;

const dfa = toDFA(regex);
dfa.complement();

console.log(toRegExp(dfa));
// => /(?:(?:[^A]|A+(?:[^AB]|B+[^B]))[^]*)?/i

Converting a JS RegExp to an NFA

In the above examples, we have been using the toNFA helper function to parse and convert RegExps. This function assumes that the given RegExp is a pure regular expression without assertions and backreferences and will throw an error if the assumption is not met.

However, the JS parser and NFA.fromRegex provide some options to work around and even solve this problem.

Backreferences

Firstly, the parser will automatically resolve simple backreferences. Even toNFA will do this since it's on by default:

console.log(toRegExp(toNFA(/("|').*?\1/)));
// => /".*"|'.*'/i

But it will throw an error for non-trivial backreferences that cannot be resolved:

toNFA(/(#+).*\1|foo/);
// Error: Backreferences are not supported.

The only way to parse the RegExp despite unresolvable backreferences is to remove the backreferences. This means that the result will be imperfect but it might still be useful.

const regex = /(#+).*\1|foo/;
const { expression } =
	JS.Parser.fromLiteral(regex).parse({ backreferences: "disable" });

console.log(JS.toLiteral(expression));
// => { source: 'foo', flags: '' }

Note that the foo alternative is kept because it is completely unaffected by the unresolvable backreferences.

Assertions

While the parser and AST format can handle assertions, the NFA construction cannot.

const regex = /\b(?!\d)\w+\b|->/;
const { expression, maxCharacter } = JS.Parser.fromLiteral(regex).parse();

console.log(JS.toLiteral(expression));
// => { source: '\\b(?!\\d)\\w+\\b|->', flags: 'i' }

NFA.fromRegex(expression, { maxCharacter });
// Error: Assertions are not supported yet.

Similarly to backreferences, we can let the parser remove them:

const regex = /\b(?!\d)\w+\b|->/;
const { expression, maxCharacter } =
	JS.Parser.fromLiteral(regex).parse({ assertions: "disable" });

console.log(JS.toLiteral(expression));
// => { source: '->', flags: 'i' }

const nfa = NFA.fromRegex(expression, { maxCharacter });
console.log(toRegExp(nfa));
// => /->/i

Or we can let the NFA construction method remove them:

const regex = /\b(?!\d)\w+\b|->/;
const { expression, maxCharacter } = JS.Parser.fromLiteral(regex).parse();

console.log(JS.toLiteral(expression));
// => { source: '\\b(?!\\d)\\w+\\b|->', flags: 'i' }

const nfa = NFA.fromRegex(expression, { maxCharacter }, { assertions: "disable" });
console.log(toRegExp(nfa));
// => /->/i

Prefer using the parser to remove assertions if possible. The parser is quite clever and will optimize based on that assertions can be removed resulting in faster parse times.

However, simply removing assertions is not ideal since they are a lot more common than backreferences. To work around this, refa has AST transformers. AST transformers can make changes to a given AST. While each transformer is rather simple, they can also work together to accomplish more complex tasks. Applying and removing assertions is one such task.

The simplest transformer to remove assertions (among other things) is the simplify transformer. It will inline expressions, remove dead branches, apply/remove assertions, optimize quantifiers, and more.

import { JS, NFA, Transformers, transform } from "refa";

const regex = /\b(?!\d)\w+\b|->/;
const { expression, maxCharacter } = JS.Parser.fromLiteral(regex).parse();
console.log(JS.toLiteral(expression));
// => { source: '\\b(?!\\d)\\w+\\b|->', flags: '' }

const modifiedExpression = transform(Transformers.simplify(), expression);
console.log(JS.toLiteral(modifiedExpression));
// => { source: '(?<!\\w)[A-Z_]\\w*(?!\\w)|->', flags: 'i' }

// Most assertions have been removed but the patterns are still equivalent.
// The only assertions left assert characters beyond the edge of the pattern.
// Removing those assertions is easy but slightly changes the pattern.

const finalExpression = transform(Transformers.patternEdgeAssertions({ remove: true }), modifiedExpression);
console.log(JS.toLiteral(finalExpression));
// => { source: '[A-Z_]\\w*|->', flags: 'i' }

const nfa = NFA.fromRegex(finalExpression, { maxCharacter });
console.log(JS.toLiteral(nfa.toRegex()));
// => { source: '->|[A-Z_]\\w*', flags: 'i' }

AST transformers can handle a lot of assertions, but there are limitations. Transformers cannot handle assertions that are too complex or require large-scale changes to the AST. Let's take a look at a few examples:

import { JS, Transformers, transform } from "refa";

function simplify(regex: RegExp): void {
  const { expression } = JS.Parser.fromLiteral(regex).parse();

  const simplifiedExpression = transform(Transformers.simplify(), expression);

  const literal = JS.toLiteral(simplifiedExpression);
  console.log(new RegExp(literal.source, literal.flags));
}

simplify(/\b(?!\d)\b\w+\b\s*\(/);
// => /(?<!\w)[A-Z_]\w*\s*\(/i
simplify(/(?:^|@)\b\w+\b/);
// => /(?:^|@)\w+(?!\w)/
simplify(/"""(?:(?!""").)*"""/s);
// => /"""(?:"{0,2}[^"])*"""/
simplify(/"""((?!""")(?:[^\\]|\\"))*"""/);
// => /"""(?:"{0,2}(?:[^"\\]|\\"))*"""/
simplify(/<title>(?:(?!<\/title>).)*<\/title>/s);
// => /<title>(?:[^<]|<+(?:[^/<]|\/(?!title>)))*<+\/title>/
simplify(/^```$.*?^```$/ms);
// => /^```[\n\r\u2028\u2029](?:[^]*?[\n\r\u2028\u2029])??```$/m

Transformers.simplify is very aggressive when it comes to assertions. It will try to remove assertions whenever possible even if it means that the overall AST will become more complex (within some limits). This may result in longer/more complex regexes, but it will also allow NFA and ENFA to support many more regexes.