npm package discovery and stats viewer.

Discover Tips

  • General search

    [free text search, go nuts!]

  • Package details

    pkg:[package-name]

  • User packages

    @[username]

Sponsor

Optimize Toolset

I’ve always been into building performant and accessible sites, but lately I’ve been taking it extremely seriously. So much so that I’ve been building a tool to help me optimize and monitor the sites that I build to make sure that I’m making an attempt to offer the best experience to those who visit them. If you’re into performant, accessible and SEO friendly sites, you might like it too! You can check it out at Optimize Toolset.

About

Hi, 👋, I’m Ryan Hefner  and I built this site for me, and you! The goal of this site was to provide an easy way for me to check the stats on my npm packages, both for prioritizing issues and updates, and to give me a little kick in the pants to keep up on stuff.

As I was building it, I realized that I was actually using the tool to build the tool, and figured I might as well put this out there and hopefully others will find it to be a fast and useful way to search and browse npm packages as I have.

If you’re interested in other things I’m working on, follow me on Twitter or check out the open source projects I’ve been publishing on GitHub.

I am also working on a Twitter bot for this site to tweet the most popular, newest, random packages from npm. Please follow that account now and it will start sending out packages soon–ish.

Open Software & Tools

This site wouldn’t be possible without the immense generosity and tireless efforts from the people who make contributions to the world and share their work via open source initiatives. Thank you 🙏

© 2026 – Pkg Stats / Ryan Hefner

hnsw

v1.1.1

Published

A TypeScript implementation of HNSW (Hierarchical Navigable Small World) algorithm for approximate nearest neighbor search

Readme

HNSW

npm version license

This is a small Typescript package that implements the Hierarchical Navigable Small Worlds algorithm for approximate nearest neighbor search.

I wrote this package because I wanted to do efficient vector search directly in the client browser. All the other implementations I found for TS were either bindings for libraries written in other languages, or dealt with WASM compilation complexity.

This is not the fastest, most fully featured, or most memory efficient implementation of HNSW. It is, however, a simple and easy to use implementation that is fast enough for many use cases.

Included is a simple persistent storage layer that uses IndexedDB to store the graph.

Installation

npm install hnsw

Usage

Ephemeral index in-memory:

import { HNSW } from 'hnsw';

// Simple example
const hnsw = new HNSW(16, 200, 5, 'cosine', 50);

// Make some data
const data = [
{id: 1, vector: [1, 2, 3, 4, 5]},
{id: 2, vector: [2, 3, 4, 5, 6]},
{id: 3, vector: [3, 4, 5, 6, 7]},
{id: 4, vector: [4, 5, 6, 7, 8]},
{id: 5, vector: [5, 6, 7, 8, 9]}
]

// Build the index
await hnsw.buildIndex(data);

// Search for nearest neighbors
const results = hnsw.searchKNN([6, 7, 8, 9, 10], 2);
const resultsWithEf = hnsw.searchKNN([6, 7, 8, 9, 10], 2, { efSearch: 100 });
console.log(results);

Persistent index using IndexedDB:

import { HNSWWithDB } from 'hnsw';

// With persistence
const index = await HNSWWithDB.create(16, 200, 'my-index', 50);

// Make some data
const data = [
{id: 1, vector: [1, 2, 3, 4, 5]},
{id: 2, vector: [2, 3, 4, 5, 6]},
{id: 3, vector: [3, 4, 5, 6, 7]},
{id: 4, vector: [4, 5, 6, 7, 8]},
{id: 5, vector: [5, 6, 7, 8, 9]}
]

// Build the index
await index.buildIndex(data);
await index.saveIndex();

// Load the same index from disk
const index2 = await HNSWWithDB.create(16, 200, 'my-index', 50);
await index2.loadIndex();

// Search for nearest neighbors
const results2 = index2.searchKNN([6, 7, 8, 9, 10], 2);
console.log(results2);

// Delete the index
await index2.deleteIndex();

Notes:

  • The metric determines how scores are computed: cosine uses cosine similarity and euclidean uses an inverse-distance similarity (higher is better in both cases).
  • efSearch controls query-time exploration and should be at least k for best recall.

API Reference

new HNSW(M, efConstruction, d?, metric?, efSearch?)

  • M: Max neighbors stored per node and layer. Higher values usually improve recall and memory cost.
  • efConstruction: Build-time exploration depth. Higher values improve index quality and build time cost.
  • d: Vector dimension. If omitted, inferred from first inserted vector.
  • metric: cosine or euclidean.
  • efSearch: Query-time exploration depth. Higher values improve recall and query latency cost.

buildIndex(data, options?)

  • data: Array of { id, vector }.
  • options.onProgress(current, total): Optional progress callback.
  • options.progressInterval: Callback cadence (default 10000).

searchKNN(query, k, options?)

  • Returns up to k results with shape { id, score }.
  • options.efSearch: Per-query override. Effective search breadth is max(k, efSearch).

toJSON() / HNSW.fromJSON(json)

  • Serialize and restore in-memory indices for transport or persistence.

HNSWWithDB.create(M, efConstruction, dbName, efSearch?)

  • Creates an IndexedDB-backed index (browser/runtime with IndexedDB support).
  • saveIndex(): Persist current graph.
  • loadIndex(): Load previously persisted graph (no-op if missing).
  • deleteIndex(): Delete persisted graph and reinitialize DB.
  • close(): Close the active IndexedDB connection.

Tuning Guide

  • Start with M=16, efConstruction=200, efSearch=50.
  • Increase efSearch first when recall is too low.
  • Increase M for tougher datasets when memory budget allows.
  • Keep efSearch >= k for better recall consistency.

Limitations

  • This implementation prioritizes simplicity over peak throughput and memory efficiency.
  • IndexedDB support depends on environment support for IndexedDB APIs.
  • Benchmark tools under src/bench are maintained as CLI utilities and are not part of the runtime API surface.

Benchmarks

A lightweight benchmark harness is available to validate recall/latency tradeoffs and the impact of parameters like efSearch, M, and efConstruction.

Build the project first:

npm run build

Download SIFT small (10k) dataset:

node dist/bench/download.js --extract

Synthetic dataset (fast sanity check):

node dist/bench/run.js --mode synthetic --count 10000 --dim 64 --metric cosine

FVECS dataset (SIFT/GloVe-style):

node dist/bench/run.js --mode fvecs --base bench/datasets/siftsmall_base.fvecs --query bench/datasets/siftsmall_query.fvecs --metric euclidean --limit 10000 --query-limit 100

Compare results (baseline vs changes):

node dist/bench/report.js --base bench/outputs/baseline.json --candidate bench/outputs/changes.json --format csv --output bench/outputs/compare.csv

One-shot compare (runs baseline + candidate + report in one command):

node dist/bench/compare.js --base-ref HEAD~1 --candidate-ref HEAD --mode fvecs --base bench/datasets/siftsmall_base.fvecs --query bench/datasets/siftsmall_query.fvecs --metric euclidean --limit 10000 --query-limit 100

For more details, see bench/README.md.