Skip to content

uhop/nano-bench

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

148 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

nano-benchmark NPM version

nano-benchmark provides command-line utilities for micro-benchmarking code with nonparametric statistics and significance testing.

Two utilities are available:

  • nano-watch — continuously benchmarks a single function, showing live statistics and memory usage.
  • nano-bench — benchmarks and compares multiple functions, calculating confidence intervals and statistical significance.

Designed for performance tuning of small, fast code snippets used in tight loops.

Visual samples

nano-watch

nano-watch

nano-bench

nano-bench

Installation

npm install nano-benchmark

Deno and Bun support

Use --self to get the script path for Deno and Bun:

npx nano-bench benchmark.js
bun `npx nano-bench --self` benchmark.js
deno run --allow-read --allow-hrtime `npx nano-bench --self` benchmark.js
deno run -A `npx nano-bench --self` benchmark.js
node `npx nano-bench --self` benchmark.js

For Deno, --allow-read is required and --allow-hrtime is recommended. Use -A for convenience in safe environments.

Documentation

With a global install (npm install -g nano-benchmark) both utilities are available by name. Otherwise, prefix with npx (e.g., npx nano-watch) or add them to your package.json scripts. Run with --help for details on arguments.

Both utilities import a module and benchmark its (default) export. nano-bench expects an object whose properties are the functions to compare. nano-watch accepts the same format or a single function.

Example module for nano-bench (bench-strings-concat.js):

export default {
  strings: n => {
    const a = 'a',
      b = 'b';
    for (let i = 0; i < n; ++i) {
      const x = a + '-' + b;
    }
  },
  backticks: n => {
    const a = 'a',
      b = 'b';
    for (let i = 0; i < n; ++i) {
      const x = `${a}-${b}`;
    }
  },
  join: n => {
    const a = 'a',
      b = 'b';
    for (let i = 0; i < n; ++i) {
      const x = [a, b].join('-');
    }
  }
};

Usage:

npx nano-bench bench-strings-concat.js
npx nano-watch bench-strings-concat.js backticks

See wiki for more details.

User Timing API integration

Pass -o / --observe to nano-bench to emit User Timing marks at calibration and sampling phase boundaries. Marks are written to the standard performance timeline and are observable via PerformanceObserver or visible in DevTools / node --inspect traces — useful for correlating benchmark variability with GC pauses, V8 optimization events, etc.

Mark / measure names follow nano-bench/<function-name>/<phase>, where phase is find-level (calibration) or series / series-par (sample collection).

import {PerformanceObserver} from 'node:perf_hooks';

const obs = new PerformanceObserver(list => {
  for (const e of list.getEntries()) {
    console.log(e.name, e.duration.toFixed(2), 'ms');
  }
});
obs.observe({entryTypes: ['measure']});

Marks have a small fixed cost per phase (no per-sample overhead), so leaving --observe on does not affect measurement accuracy. Default is off.

Library users can opt in directly: findLevel / benchmarkSeries / benchmarkSeriesPar / measure / measurePar all accept an observe option (boolean | string) — false / unset for no marks, true for the default label, or a string for a custom label.

AI agents and contributing

AI agents and AI-assisted developers: read AGENTS.md first for project rules and conventions.

Other useful files:

License

BSD 3-Clause License

Release history

  • 1.0.16: Added User Timing API integration: --observe flag.
  • 1.0.15: Updated dependencies.
  • 1.0.14: Fixed Kruskal-Wallis post-hoc (Conover-Iman) pairwise comparison bug: corrected rank variance computation and critical value distribution. Added regression test.
  • 1.0.13: Improved CLI help texts and documentation for brevity and clarity.
  • 1.0.12: Added AI coding skills for writing benchmark files (write-bench, write-watch), shipped via npm. Added findLevel() tests. Expanded test suite.
  • 1.0.11: Fixed MedianCounter.clone() bug, expanded test suite (204 tests), added CodeQL workflow, multi-OS CI matrix, and new Windsurf workflows.
  • 1.0.10: Added Prettier lint scripts, GitHub issue templates, Copilot instructions, and Windsurf workflows.
  • 1.0.9: Updated dependencies.
  • 1.0.8: Updated dependencies.
  • 1.0.7: Updated dependencies.
  • 1.0.6: Updated dependencies.
  • 1.0.5: Updated dependencies.
  • 1.0.4: Updated dependencies + added more tests.
  • 1.0.3: Updated dependencies.
  • 1.0.2: Added the --self option.
  • 1.0.1: Added "self" argument to utilities so it can be used with Deno, Bun, etc.
  • 1.0.0: Initial release.

The full release notes are in the wiki: Release notes.

Sponsor this project

  •  

Contributors