npm package discovery and stats viewer.

Discover Tips

  • General search

    [free text search, go nuts!]

  • Package details

    pkg:[package-name]

  • User packages

    @[username]

Sponsor

Optimize Toolset

I’ve always been into building performant and accessible sites, but lately I’ve been taking it extremely seriously. So much so that I’ve been building a tool to help me optimize and monitor the sites that I build to make sure that I’m making an attempt to offer the best experience to those who visit them. If you’re into performant, accessible and SEO friendly sites, you might like it too! You can check it out at Optimize Toolset.

About

Hi, 👋, I’m Ryan Hefner  and I built this site for me, and you! The goal of this site was to provide an easy way for me to check the stats on my npm packages, both for prioritizing issues and updates, and to give me a little kick in the pants to keep up on stuff.

As I was building it, I realized that I was actually using the tool to build the tool, and figured I might as well put this out there and hopefully others will find it to be a fast and useful way to search and browse npm packages as I have.

If you’re interested in other things I’m working on, follow me on Twitter or check out the open source projects I’ve been publishing on GitHub.

I am also working on a Twitter bot for this site to tweet the most popular, newest, random packages from npm. Please follow that account now and it will start sending out packages soon–ish.

Open Software & Tools

This site wouldn’t be possible without the immense generosity and tireless efforts from the people who make contributions to the world and share their work via open source initiatives. Thank you 🙏

© 2024 – Pkg Stats / Ryan Hefner

cf-deps

v2.0.0

Published

cli tool to generate and publish CloudFormation dependencies from yml files

Downloads

554

Readme

Gitpod Ready-to-Code

cf-deps (and cf-groups)

Contains 2 CLI binaries:

  1. cf-deps: Publishes a list of serverless dependencies on CloudFormation stacks to a centralized repo
  2. cf-groups: Consumes the published dependency records from cf-deps and generates deploy groups

Why?

  • To allow you to scale n-number of serverless stacks with minmal effort on tracking deploy dependencies
  • Shed light on areas you might have too coupled

Quick Setup

Install

npm i -g cf-deps

Get help

cf-deps help
cf-groups help

Tips

  • Use environment variables for configuring options so that both cf-deps and cf-groups use the same settings with minimal effort
  • Configure cf-deps to run automatically on master (or trunk) for all your related projects in CI/CD (cf-groups is better for on-demand local use)
  • Use the --save or -s flag with cf-groups to save a local file of deploy groups as a markdown file for "pretty" viewing in a markdown tool; combine this with the option for link pattern to generate clickable links to your repos
  • AVOID depending on too many CloudFormation stacks; These tools aid managing where you do, but make every effort to build independent and decoupled systems
  • Try this out with the test git server in the source code

cf-deps

  • Finds cf refs (${cf:stackName.output}) from yaml (or yml) files
  • Makes a list of stackNames (removes duplicates)
  • Commit list of stackNames to git repo for aggregated storage of stack dependencies

Sample stack dependency json:

{
    "stack-name-a": [],
    "stack-name-b": [
        "stack-name-a"
    ],
    "stack-name-c": [
        "stack-name-b"
    ],
    "stack-name-d": [
        "stack-name-a",
        "stack-name-c"
    ]
}

If you build a new stack, say "stack-name-e", and run cf-deps in it's respective CI/CD build, then it would add "stack-name-e": [ /* list of cf stackName refs */ ] to the json sample above. From the living centralized dependency json, you have a quick one-stop-shop for deployment dependencies. Further automation can be built on top of this to generate deploy groups in order of their dependencies.

Usage

Be sure to checkout cf-deps help

With ENV Variables (Useful for CI/CD with n-number of stacks)

export CF_DEPS_JSON_FILE="<json file to store deps in>"
export CF_DEPS_REMOTE_GIT="<git clone url for storage repo>"
export CF_DEPS_GIT_USR_NAME="<git user.name>"
export CF_DEPS_GIT_USR_MAIL="<git user.email>"
cf-deps

With CLI Options

cf-deps \
  -f <json file to store deps in> \
  -r <git clone url for storage repo> \
  -u <git user.name> \
  -m <git user.email>

Known Limitations

Sometimes your projects can have complex CloudFormation references that the RegExp cannot account for; An example, a cf ref to itself in a different region, like so:

${cf.${self:custom.replicationRegion}:${self:service}-${opt:stage}.BucketName}

In this case, a stackName, "custom" would be collected by cf-deps, but this isn't realistic. Instead, you can add "custom" as an excludes. See cf-deps help.

Alternatively, you could do some processing before cf-deps to

  1. Escape ${cf lookups
  2. Serverless print evaluated config
  3. Unescape ${cf lookups
  4. (Optional) remove defaults from ${cf..., default}
  5. Run cf-deps

Example

# 1. escape cf lookups
sed -i "s/\${cf/cfn{cf/g" $(git grep -l "\${cf")
# 2. serverless print evaluated config
npx sls print -s my-stage -r my-region 2>/dev/null > sls-config.yml
# 3. unescape cf lookups
sed -i "s/cfn{cf/\${cf/g" sls-config.yml $(git grep -l "cfn{cf")
# 4. remove defaults from cf lookups
grep -o '\${.*,.*}' sls-config.yml | sed 's/,.*}/}/g' >> sls-config.yml

# 5. run cf-deps
export CF_DEPS_FILENAME_PTN="sls-config.yml"
cf-deps

cf-groups

  • Clones the latest commit on the repo branch specified; should be the storage repo (and json file) specified for cf-deps
  • Generates ordered deploy groups based on dependencies provided in the storage repo
  • Can include all recorded stacks based on dependencies or a subset provided by options
  • Creates a list of missed dependencies if deploy set does not include them all
  • Has an option to specify a custom filter for dependencies (in case you have stacks that do not record dependencies in yml/yaml files in the ${cf:stackName.output} format)

Warning: I don't think this handles circular dependencies; I think it errors out, stack overflow or something with the recursive logic. Haven't had the "pleasure" to test that unhappy path.

Sample Dependencies Filter

  • Store the custom filter as a *.js file in the storage repo
module.exports = (repoName, deps) => {
  if (repoName !== 'account-resources') {
    // such as a stack that provisions foundational
    // things that must come before anything else
    deps.push('account-resources');
  }
  if (repoName.endsWith('-client')
    || [
      'hello-world-app',
      'idk-poc',
      'tech-debt-thing',
    ].includes(repoName)) {
        
    deps.push('shared-client-resources');
    deps.push('some-important-server');
  }
};

Usage

Be sure to checkout cf-groups help

With ENV Variables

export CF_DEPS_JSON_FILE="cf-deps.json";
export CF_DEPS_REMOTE_GIT="[email protected]/repo.git";
cf-groups

# with a filter
cf-groups --filter my-custom-filter.js

With CLI Options

cf-groups \
  -f cf-deps.json \
  -r [email protected]/repo.git

# wilt a filter
cf-groups \
  -f cf-deps.json \
  -r [email protected]/repo.git \
  --filter my-custom-filter.js

Working with the Source Code

Install the things and link cf-deps for CLI use:

npm ci
npm link

Local Git Server

This source code includes a simple local git server (requires docker and docker-compose) which can make testing easier and faster.

Build and Run

  1. Build the needful
    docker-compose build
  2. Run the needful
    docker-compose up

With the docker-compose stack running, you can clone the repo(s) that are in the local git server

cd /tmp && git clone git://localhost/cf-deps-store

On startup (aka, during docker-compose up), the git-server initiates git repos from the folders in the test/git-server/seed-repos. You can then clone the repo(s), change, commit, and push them as needed while the server is running. Restart the server and start over from the seed-state.

Stop/Start

Currently, the git-server cannot handle start/stop of the container, so you have to make sure to remove the container before starting again; simply run:

docker-compose down

Then you can run the needful again:

docker-compose up