npm package discovery and stats viewer.

Discover Tips

  • General search

    [free text search, go nuts!]

  • Package details

    pkg:[package-name]

  • User packages

    @[username]

Sponsor

Optimize Toolset

I’ve always been into building performant and accessible sites, but lately I’ve been taking it extremely seriously. So much so that I’ve been building a tool to help me optimize and monitor the sites that I build to make sure that I’m making an attempt to offer the best experience to those who visit them. If you’re into performant, accessible and SEO friendly sites, you might like it too! You can check it out at Optimize Toolset.

About

Hi, 👋, I’m Ryan Hefner  and I built this site for me, and you! The goal of this site was to provide an easy way for me to check the stats on my npm packages, both for prioritizing issues and updates, and to give me a little kick in the pants to keep up on stuff.

As I was building it, I realized that I was actually using the tool to build the tool, and figured I might as well put this out there and hopefully others will find it to be a fast and useful way to search and browse npm packages as I have.

If you’re interested in other things I’m working on, follow me on Twitter or check out the open source projects I’ve been publishing on GitHub.

I am also working on a Twitter bot for this site to tweet the most popular, newest, random packages from npm. Please follow that account now and it will start sending out packages soon–ish.

Open Software & Tools

This site wouldn’t be possible without the immense generosity and tireless efforts from the people who make contributions to the world and share their work via open source initiatives. Thank you 🙏

© 2024 – Pkg Stats / Ryan Hefner

@jrc03c/filedb

v0.0.24

Published

a database solution that reads and writes plain text files from disk

Downloads

50

Readme

Intro

FileDB is a simple database tool for Node. It writes and reads plaintext files to and from disk.

Installation

npm install --save @jrc03c/filedb

Usage

const FileDB = require("@jrc03c/filedb")

// create a new database, specifying the directory in which to write data
// (which will be created automatically if it doesn't exist)
const db = new FileDB("some/directory")

// write data
const payload = { x: 5, y: 7, z: "Hello, world!" }
db.writeSync("/foo", payload)

// read data
const results = db.readSync("/foo")
console.log(results)
// {x: 5, y: 7, z: "Hello, world!"}

// delete data
db.writeSync("/foo", null)

FileDB works by writing data as files and folders on disk. There are a few implications of this design:

  • It's possible to add new stuff to the "database" (the FileDB instance's working directory) without using FileDB. In other words, you can manually modify the files and folders in the database without using filedb at all!
  • It's easy to read and understand what data is being written and read because you can examine and modify the files and folders directly.

An example in detail

Suppose that Alice is working on a project that needs a database. Her project is located /home/alice/my-project. She decides she needs a database for the project, and she wants the data to live in /home/alice/my-project/data. She can create the new filedb database in one of two ways:

  1. She can pass the absolute path to the database directory into the FileDB constructor. For example: new FileDB("/home/alice/my-project/data")
  2. From anywhere along the path to the database directory, she can pass a relative path into the FileDB constructor. For example, from within the /home/alice/my-project directory, she can do: new FileDB("data")

So, Alice creates a script at /home/alice/my-project/index.js with this content:

const FileDB = require("@jrc03c/filedb")
const db = new FileDB("data")

Now she wants to write some data into the database. Suppose that her project involves keeping track of some user data. She could do something like this (where the long, random-looking strings are the MD5 hashes of the users' email addresses):

db.writeSync("/users", {
  "c160f8cc69a4f0bf2b0362752353d060": {
    name: "Alice",
    email: "[email protected]",
    age: 21,
    preferences: {
      darkMode: true,
    },
  },
  "4b9bb80620f03eb3719e0a061c14283d": {
    name: "Bob",
    email: "[email protected]",
    age: 34,
    preferences: {
      darkMode: false,
    },
  },
  "f6f24749c91477e8c31c26b21a51c732": {
    name: "Candice",
    email: "[email protected]",
    age: 56,
    preferences: {
      darkMode: true,
    },
  },
})

Now the structure of her /home/alice/my-project/data directory looks like this:

data
└── users
    ├── 4b9bb80620f03eb3719e0a061c14283d
    │   ├── age
    │   ├── email
    │   ├── name
    │   └── preferences
    │       └── darkMode
    ├── c160f8cc69a4f0bf2b0362752353d060
    │   ├── age
    │   ├── email
    │   ├── name
    │   └── preferences
    │       └── darkMode
    └── f6f24749c91477e8c31c26b21a51c732
        ├── age
        ├── email
        ├── name
        └── preferences
            └── darkMode

💡 NOTE: The order in which the users are listed above is different from the order in which the users were listed in the object that Alice stored in the database. That's because the tree command I used to print out this fancy directory tree sorts items before printing them, and Bob's email hash ("4b9bb80620f03eb3719e0a061c14283d") comes before Alice's email hash ("c160f8cc69a4f0bf2b0362752353d060") when sorted in this way. We'll see the same thing in a moment when Alice reads the data back from disk.

Each of the leaf nodes on this tree are files containing the value that corresponds to that key name. For example:

cat /home/alice/my-project/data/users/c160f8cc69a4f0bf2b0362752353d060/name
# "Alice"

To retrieve the user data later, Alice can do this:

const myData = db.readSync("/users")
console.log(myData)

// {
//   '4b9bb80620f03eb3719e0a061c14283d': {
//     age: 34,
//     email: '[email protected]',
//     name: 'Bob',
//     preferences: { darkMode: false }
//   },
//   c160f8cc69a4f0bf2b0362752353d060: {
//     age: 21,
//     email: '[email protected]',
//     name: 'Alice',
//     preferences: { darkMode: true }
//   },
//   f6f24749c91477e8c31c26b21a51c732: {
//     age: 56,
//     email: '[email protected]',
//     name: 'Candice',
//     preferences: { darkMode: true }
//   }
// }

She can also retrieve specific bits of data by reading further along each path. For example, she can retrieve her own dark mode preference this way:

const aliceDarkModePreference = db.readSync(
  "/users/c160f8cc69a4f0bf2b0362752353d060/preferences/darkMode"
)

console.log(aliceDarkModePreference)
// true

Reading and writing arrays

There's only one case where filedb writes extra data in addition to what you pass to it: when it needs to write an array to disk. Suppose that Alice wants to store a list of email addresses in the database for quick reference. Should could do something like this:

db.writeSync("/emails", [
  "[email protected]",
  "[email protected]",
  "[email protected]",
])

Now, her directory tree looks like this:

data
├── emails
│   ├── 0
│   ├── 1
│   ├── 2
│   └── .filedb.meta.is-array
└── users
    ├── 4b9bb80620f03eb3719e0a061c14283d
    │   ├── age
    │   ├── email
    │   ├── name
    │   └── preferences
    │       └── darkMode
    ├── c160f8cc69a4f0bf2b0362752353d060
    │   ├── age
    │   ├── email
    │   ├── name
    │   └── preferences
    │       └── darkMode
    └── f6f24749c91477e8c31c26b21a51c732
        ├── age
        ├── email
        ├── name
        └── preferences
            └── darkMode

You may have spotted an extra file in there called .filedb.meta.is-array under data/emails. That file is just a little flag that filedb uses to remind itself that the data stored in the data/users directory is an array. Right now, if Alice reads the data back from disk, she gets what she expects:

const emailsArray = db.readSync("/emails")
console.log(emailsArray)
// [ '[email protected]', '[email protected]', '[email protected]' ]

However, if she deletes the .filedb.meta.is-array file, then there's a chance that the data can be read back as an object. If all of the keys are whole numbers over a range in which no numbers are skipped (e.g., [0, 1, 2, 3, 4, 5, 6]), then the data will be read back as an array. But if any of the keys are not whole numbers, or if any whole numbers are missing from the range of whole numbers (e.g., if 2 is missing from [0, 1, 3, 4, 5, 6]), then the data will be read back as an object. We can see this if Alice intentionally deletes both the .filedb.meta.is-array file and the data at index 1:

// delete those files
db.writeSync("/emails/.filedb.meta.is-array", null)
db.writeSync("/emails/1", null)

// read the data
const emailsObject = db.readSync("/emails")
console.log(emailsObject)
// { '0': '[email protected]', '2': '[email protected]' }

If Alice deletes the data at index 1 but does not delete the .filedb.meta.is-array file, then the data will be read back as an array with a missing value at index 1.

Valid filesystem paths

Since every index and key down through an object or array is converted to a file or directory and written to disk, every index and key must be safe to use as a name for a file or directory. Since various filesystems apparently allow different sets of characters, I've set the list of valid characters to be those that seem common to all Unix filesystems: forward slashes ("/"), alphanumerics (0-9, A-Z, a-z), hyphens ("-"), underscores ("_"), and periods ("."). Paths also cannot contain whitespace, and they cannot include file or directory names that consist only of a single period (".") or a double period ("..").

For example, these won't work:

// Nope. 🔴 😿
db.readSync("../something")
db.readSync("path/./to/./something")
db.readSync("my spaced out path")
db.readSync("$HOME")
db.readSync("`pwd`")

But these will:

// Yep! 🟢 😸
db.readSync("/my/cool/path")
db.readSync("this_is-totally.fine")
db.readSync(".so.is.this.")
db.readSync("1/2/3/4")

Also note that paths are subject to whatever other constraints are placed on them by the filesystem, including maximum path lengths. For example, you might find that db.writeSync("/smol", true) throws an error if the absolute path to /smol is too long.

Finally, note that paths ending in a forward slash ("/") don't make any sense because they're neither a file nor a directory; so all such trailing forward slashes are merely removed, no warnings are given, and no errors are thrown.

Maximum read depth

Data can become quite deeply nested, and it's not always necessary to retrieve all of it. In such cases, a maximum read depth can be passed to the read or readSync methods.

Suppose that Alice wants to get a list of all of the MD5 hashes of her users' email addresses. She can accomplish this by reading the /users path but restricting the depth from which data is read. For example:

const hashes = db.readSync("/users", 0)
console.log(hashes)

// [
//   '4b9bb80620f03eb3719e0a061c14283d',
//   'c160f8cc69a4f0bf2b0362752353d060',
//   'f6f24749c91477e8c31c26b21a51c732'
// ]

Alice's user data has 4 levels of depth: users → email hash → preferences → dark mode. Reading the full depth of the data is equivalent to passing a 3 as the maximum read depth (since the depth starts at 0):

// no maximum depth
console.log(db.readSync("/users"))

// {
//   '4b9bb80620f03eb3719e0a061c14283d': {
//     age: 34,
//     email: '[email protected]',
//     name: 'Bob',
//     preferences: { darkMode: false }
//   },
//   c160f8cc69a4f0bf2b0362752353d060: {
//     age: 21,
//     email: '[email protected]',
//     name: 'Alice',
//     preferences: { darkMode: true }
//   },
//   f6f24749c91477e8c31c26b21a51c732: {
//     age: 56,
//     email: '[email protected]',
//     name: 'Candice',
//     preferences: { darkMode: true }
//   }
// }

// maximum depth of 3, which in this case is the same as having no maximum depth
console.log(db.readSync("/users", 3))

// {
//   '4b9bb80620f03eb3719e0a061c14283d': {
//     age: 34,
//     email: '[email protected]',
//     name: 'Bob',
//     preferences: { darkMode: false }
//   },
//   c160f8cc69a4f0bf2b0362752353d060: {
//     age: 21,
//     email: '[email protected]',
//     name: 'Alice',
//     preferences: { darkMode: true }
//   },
//   f6f24749c91477e8c31c26b21a51c732: {
//     age: 56,
//     email: '[email protected]',
//     name: 'Candice',
//     preferences: { darkMode: true }
//   }
// }

API

FileDB([path])

Constructs a new FileDB instance. The given path (which is optional) is the directory in which the instance will do all of its reading and writing. Paths can be relative or absolute. If no path is provided, then the path resolves to the current working directory.

exists(key, [callback])

Asynchronously checks to see whether or not a key exists. Returns a Promise that resolves to a boolean. Passing a callback function is optional.

existsSync(key)

Synchronously checks to see whether or not a key exists. Returns a boolean.

write(key, value, [ignored], [callback])

Asynchronously writes a key-value pair to disk. Returns a Promise that resolves to a boolean indicating whether or not the value was written to disk. Generally, this will be a true value, but it can be false in cases where the given key matches a pattern in the ignored list. Passing an ignored list is optional; but if passed, it must be an array containing strings and/or regular expressions against which to match paths. Passing a callback function is optional.

writeSync(key, value, [ignored])

Synchronously writes a key-value pair to disk. Returns a boolean indicating whether or not the value was written to disk. Generally, this will be a true value, but it can be false in cases where the given key matches a pattern in the ignored list. Passing an ignored list is optional; but if passed, it must be an array containing strings and/or regular expressions against which to match paths.

read(key, [maxDepth], [ignored], [callback])

Asynchronously reads the value of a key from disk. Returns a Promise that resolves to whatever kind of data the stored value represents. Passing a callback function is optional. A maximum depth value is also optional and represents the depth of data to return. For example, if there's a key at "/a/b/c/d/e/.../z", and I call db.read("/a/b/c", 3), then the returned data will only go as deep as /a/b/c. See the example below for further clarification. Passing an ignored list is optional; but if passed, it must be an array containing strings and/or regular expressions against which to match paths.

readSync(key, [maxDepth], [ignored])

Synchronously reads the value of a key from disk. Returns whatever kind of data the stored value represents. A maximum depth value is optional and represents the depth of data to return. For example, if there's a key at "/a/b/c/d/e/.../z", and I call db.read("/a/b/c", 3), then the returned data will only go as deep as /a/b/c. See the example below for further clarification. Passing an ignored list is optional; but if passed, it must be an array containing strings and/or regular expressions against which to match paths.

Caveats and other musings

  • I've tried to mitigate the risk of accidentally (or intentionally) using a key that converts to a path outside of the database's working directory. Valid filesystem paths are described above. However, it may still be possible to work around this roadblock. If you find a way to do so, please let me know so I can patch it up! If you're really worried about this, though, one potential solution could be to create a password and symmetrically encrypt each key with it such that the key becomes an alphanumeric string before you try to read or write with it. Thus if a key somehow contains malicious code that could allow it to read from or write to parts of the filesystem outside of the database's working directory, then this method would convert the key to only alphanumeric characters for storage; but it could be decrypted later using the same password. Note that this is absolutely not a method for keeping data safe or private; it's only a possible strategy for mitigating nefarious reads and writes caused by a nefarious key. In fact, I'd recommend finding a super simple encryption algorithm, one that can work fast and produce relatively short strings, since the goal in this particular case has nothing to do with encrypting the data at rest.
  • This package intentionally doesn't include any options to encrypt the data at rest because such a feature is out of scope.
  • I don't know what I'm doing. This is probably horribly designed and executed. But it works for my needs, so I'm happy with it.
  • I have no idea how it'll respond to binary data. I haven't tried it yet. But I suspect that, as long as the data can be stringified, it can be written and read with FileDB. I should probably add that to the to-do list...
  • Good luck! Feel free to give me feedback!