npm package discovery and stats viewer.

Discover Tips

  • General search

    [free text search, go nuts!]

  • Package details

    pkg:[package-name]

  • User packages

    @[username]

Sponsor

Optimize Toolset

I’ve always been into building performant and accessible sites, but lately I’ve been taking it extremely seriously. So much so that I’ve been building a tool to help me optimize and monitor the sites that I build to make sure that I’m making an attempt to offer the best experience to those who visit them. If you’re into performant, accessible and SEO friendly sites, you might like it too! You can check it out at Optimize Toolset.

About

Hi, 👋, I’m Ryan Hefner  and I built this site for me, and you! The goal of this site was to provide an easy way for me to check the stats on my npm packages, both for prioritizing issues and updates, and to give me a little kick in the pants to keep up on stuff.

As I was building it, I realized that I was actually using the tool to build the tool, and figured I might as well put this out there and hopefully others will find it to be a fast and useful way to search and browse npm packages as I have.

If you’re interested in other things I’m working on, follow me on Twitter or check out the open source projects I’ve been publishing on GitHub.

I am also working on a Twitter bot for this site to tweet the most popular, newest, random packages from npm. Please follow that account now and it will start sending out packages soon–ish.

Open Software & Tools

This site wouldn’t be possible without the immense generosity and tireless efforts from the people who make contributions to the world and share their work via open source initiatives. Thank you 🙏

© 2024 – Pkg Stats / Ryan Hefner

art-ery

v1.120.3

Published

A pipelined business-logic framework for cloud-backed data. ArtEry conceptially allows you to write apps 100% as client-code, but with the security and performance of cloud-code.

Downloads

116

Readme

ArtEry - Client Development > Cloud Deployment

ArtEry2 is coming! Improved JavaScript usability. TypeScript support. Support for multiple pipeline groups both clientside and serverside allowing defining and using multiple APIs all within the ArtEry system. ArtEry2 also has numerous simplifications resulting from breaking functionality out into many smaller, single-purpose packages.

ArtEry is a full-stack solution for managing database-backed data. It allows you to develop, test and deploy applications using a single codebase that runs on both client and server. For the relatively small effort of defining your database "pipelines" (i.e. models) using a streamlined runtime-declarative system, you get the following "for free" - no additoinal code or configuration needed:

  1. An API server with OpenAPI docs (GraphQL possibly coming in the future)
  2. Database schema generation and migrations (using your favorate ORM as a plugin - DynanoDb supported now; Prisma coming soon)
  3. JavaScript clients for seamless integration with front-end frameworks like ReactJS with all client-side data management taken care of for you

Server-Side Developer

ArtEry allows server-side code to be written 100% declaratively through database Pipelines & Filters. No servers, routers or other infrastructure code needs to be written. ArtEry is an excellent way to express your database business logic.

Full-Stack Developer

ArtEry conceptually allows you to develop, test and debug applications 100% as client-side-code, but with the security and performance of cloud-code. When you deploy to production, it's trivial to control which code goes in the Client, Server or Both. The key thing is you can develop with the whole stack in one runtime. This means errors and tracing all happen in one place. You can see full stack-traces across your entire application within your browser's console.

In an masterstoke of uniformity, the ArtEry JavaScript API for manipulating your data via your Pipelines is identical clientside and serverside.

Benefits

  • Fastest possible development (testing, debugging, build cycle)
  • Security
  • Performance
  • Eliminate code duplication
  • Together with ArtEryServer, generate a full REST API automatically

Fastest Possible Development

Client-code has many advantages over cloud-code:

  • easier to test
  • easier to debug
  • dramatically shorter build cycle

In short, it's much faster to develop.

The key observation is code is easier to develop, test and debug when it's all in one runtime. Stack traces span your full stack. You can hot-reload your full stack for rapid development.

Security and Performance

There are some things that can't be done safely client-side:

  • Authentication
  • Authorization
  • Validation
  • TimeStamps
  • Update Counts

Pipeline filters make it easy to apply any security rules you need.

Performance

Some requests are more efficient to process in the cloud:

  • requests with require multiple cloud-side requests
    • client-to-cloud requests typically cost much more than cloud-to-cloud request
  • requests which consume a bunch of data, reduce it, and output much less data
    • cloud-to-client data-transfer typically costs much more than cloud-to-cloud
  • requests with a lot of computation

Adding custom handler-types for complex requests make this easy to do. A robust "sub-request" system makes synthesizing complex requests easy in a scalable, performant way.

Eliminate Code Duplication

Some code should run both on the cloud and client. Specifically, validation should happen client-side for the fastest possible response to client actions, /but it needs to be verified in the cloud for security. ArtEry makes it trivial to re-use code across your full stack.

Example

This is a complete example.

# filename: Post.caf
import &ArtStandardLib, &ArtEry

# simple in-memory CRUD data-store
class Post extends Pipeline

  constructor: -> @data = {}

  # crud-api
  @handlers
    get:    ({key})       -> @data[key]
    create: ({data})      -> key = randomString(); @data[key] = merge data, id: key
    update: ({key, data}) -> @data[key] = merge @data[key], data
    delete: ({key})       -> delete @data[key]

  @publicRequestTypes :get :create :update :delete

  # text-trimming filter
  @filter before:
    update: trimText = (request) -> request.withMergedData text: request.data.text?.trim()
    create: trimText

Use:

{post} = &Post
&ArtConfig.configure()

post.create data: text: "   Hello world!    "
.then ({id})   -> post.get id
.then ({text}) -> console.log text  # Hello world!

How it Works

  • Declare one or more pipelines with handlers and filters
  • Make client-side requests to those pipelines
  • The pipeline takes care of routing the request wherever it needs to go.

All requests are handled with promises in a (nearly) pure-functional way. Each filter and handler returns a NEW request or response, passing the information down the pipeline. Since each step of the pipeline is handled as a promise, you can get maximum scalable performance out of node.js.

Pipelines

Pipelines are the main structural unit for ArtEry. A pipeline consists of:

  • name: - derived from the pipeline's class-name
  • handlers: a map from request-types to handlers: [<String>]: <Handler>
  • filters:

A pipeline is a named grouping of request-types and filters. When designing an API for a database-backed backend, it's usually best to have one pipeline per table in your database.

Handlers

Handlers are just functions:

(<Request>) -> <[optional Promise] Response, null, plain-Object, or other response-compatible-data-types>

Requests and Responses

At their most simple, requests and responses both look like this:

type: <String>
props:
  key:  <any, but usually String>
  data: <any, but usually Object>
  <any, but usually nothing>

For convenience, there are some common getters:

# getters
key:  -> @props.key
data: -> @props.data

Filters

At their simplest, functions are almost exactly the same as handlers:

# before-handler
(<Request>) -> <[optional Promise] Request, Response, plain-Object, or other response-compatible-data-types>

# after-handler
(<Response>) -> <[optional Promise] Response, plain-Object, or other response-compatible-data-types>

In general, each filter applies to the whole pipeline. It can filter any before and after any request-type. In practice, you'll write filters which only filter certain request-types and perhaps only before or after the handler.

Filters can also be configured to run client-side, server-side or both:

class MyFilter extends Filter

  @location :client # :server or :both

  @before create: (request) -> ...
  @after  create: (request) -> ...

FAQ

  • Should I put my request-fields in request.props or request.data?

    • NOTE: request.data == request.props.data - in other words, data is actually a props-field with a convenient accessor.
    • You can put anything in props.data and almost anything in props itself.
    • Recommendation: If the field is one of the defined @fields, it goes in data. Else, put it in props.
    • Why? To make filters as re-usable as possible, they need to make assumptions about your request-props and response-props.
  • What do you mean by "Nearly" pure-functional?

    • To facilitate more complex sub-request patterns, each root-request has a mutable context object for shared state across all sub-requests. This is useful for functionality such as caching sub-requests. Keeping all context attached to the root request rather than somewhere else ensures we can have any number of request in flight simultaneously. In general, only advanced filter-developers should touch the context object. Almost everything you'll ever need to do can be done in a pure-functional way with normal filters and handlers.

TODO

Use the Authorization: Bearer HTTP header to pass a session on GET requests. https://stackoverflow.com/questions/33265812/best-http-authorization-header-type-for-jwt

Filters

  • instead of "filterFailures" - perhaps we should have an afterFailure path separate from after?