npm package discovery and stats viewer.

Discover Tips

  • General search

    [free text search, go nuts!]

  • Package details

    pkg:[package-name]

  • User packages

    @[username]

Sponsor

Optimize Toolset

I’ve always been into building performant and accessible sites, but lately I’ve been taking it extremely seriously. So much so that I’ve been building a tool to help me optimize and monitor the sites that I build to make sure that I’m making an attempt to offer the best experience to those who visit them. If you’re into performant, accessible and SEO friendly sites, you might like it too! You can check it out at Optimize Toolset.

About

Hi, 👋, I’m Ryan Hefner  and I built this site for me, and you! The goal of this site was to provide an easy way for me to check the stats on my npm packages, both for prioritizing issues and updates, and to give me a little kick in the pants to keep up on stuff.

As I was building it, I realized that I was actually using the tool to build the tool, and figured I might as well put this out there and hopefully others will find it to be a fast and useful way to search and browse npm packages as I have.

If you’re interested in other things I’m working on, follow me on Twitter or check out the open source projects I’ve been publishing on GitHub.

I am also working on a Twitter bot for this site to tweet the most popular, newest, random packages from npm. Please follow that account now and it will start sending out packages soon–ish.

Open Software & Tools

This site wouldn’t be possible without the immense generosity and tireless efforts from the people who make contributions to the world and share their work via open source initiatives. Thank you 🙏

© 2024 – Pkg Stats / Ryan Hefner

wu-wei-benchmarking-toolkit

v0.3.2

Published

Wu-Wei (無爲) Benchmarking Toolkit ==========================

Downloads

8

Readme

Wu-Wei (無爲) Benchmarking Toolkit

Join the chat at https://gitter.im/Sable/wu-wei-benchmarking-toolkit Build Status

Wu-Wei (non-effort) is a benchmarking toolkit developed in the Sable Lab at McGill University with the objective of simplifying the study of the performance of programming languages implementations and tools.

We aim to make the toolkit and the benchmark suites built with it:

  1. Consistent and Correct by supporting correctness checks for every language implementation of benchmarks that automatically ensure that the computation result of the benchmarks are consistent across all language implementations and correct with regard to the algorithm for known inputs;
  2. Extensible across numerical languages, benchmarks, compilers, run-time environments;
  3. Friendly to language implementation research by automating all tasks for compiler and virtual-machine research and encouraging a writing style for benchmarks that factors the core computation from the runners to minimize the non-core functions necessary to validate the output of compilers;
  4. Easy to use by automating the deployment of benchmarks, their test on virtual (web browser and others) and native platforms, as well as the gathering and reporting of relative performance data;
  5. Fast by making the setup (data generation and loading) and teardown as quick as possible so that most of the time is spent in the core computation in every language;
  6. Small by minimizing the amount of data needed to use the suite;
  7. Simple by minimizing the amount of external dependencies and tools required to run the suite;

Dependencies

Although we tried our best to minimize external dependencies, the toolkit still depends on the following external tools:

  1. Node.js
  2. Python

Individual artifacts may have more dependencies. Refer to their documentation for more details.

Getting Started

Please read the handbook for more details on how to use it.

Copyright and License

Copyright (c) 2016, Erick Lavoie, Hanfeng Chen