npm package discovery and stats viewer.

Discover Tips

  • General search

    [free text search, go nuts!]

  • Package details

    pkg:[package-name]

  • User packages

    @[username]

Sponsor

Optimize Toolset

I’ve always been into building performant and accessible sites, but lately I’ve been taking it extremely seriously. So much so that I’ve been building a tool to help me optimize and monitor the sites that I build to make sure that I’m making an attempt to offer the best experience to those who visit them. If you’re into performant, accessible and SEO friendly sites, you might like it too! You can check it out at Optimize Toolset.

About

Hi, 👋, I’m Ryan Hefner  and I built this site for me, and you! The goal of this site was to provide an easy way for me to check the stats on my npm packages, both for prioritizing issues and updates, and to give me a little kick in the pants to keep up on stuff.

As I was building it, I realized that I was actually using the tool to build the tool, and figured I might as well put this out there and hopefully others will find it to be a fast and useful way to search and browse npm packages as I have.

If you’re interested in other things I’m working on, follow me on Twitter or check out the open source projects I’ve been publishing on GitHub.

I am also working on a Twitter bot for this site to tweet the most popular, newest, random packages from npm. Please follow that account now and it will start sending out packages soon–ish.

Open Software & Tools

This site wouldn’t be possible without the immense generosity and tireless efforts from the people who make contributions to the world and share their work via open source initiatives. Thank you 🙏

© 2026 – Pkg Stats / Ryan Hefner

@mfukala/wsfs

v0.1.18

Published

wsfs: local-first virtual filesystem with IndexedDB cache and pluggable backend persistence.

Readme

wsfs

wsfs (Web Sync File System) is a browser-friendly, local-first virtual filesystem for apps that need offline sync, IndexedDB caching, optimistic concurrency (ETag/If-Match), and pluggable backend persistence—zero runtime deps, works with any REST-ish API.

What it solves (at a glance)

  • Offline-first reads/writes with IndexedDB caching and an in-memory fallback (runs in browsers, workers, and Node.js)
  • Incremental sync with watermarks + bulk reads to avoid walking the full tree; conflict events when remote ETags differ
  • Auth + partition hooks to pass proof headers/body fields through to authorize and route multi-tenant traffic to the right adapter
  • Pluggable codecs to encrypt/compress payloads before storage or the network; binary payloads travel as base64
  • Server toolkit with memory + driver-agnostic SQL persistence, pluggable into Express/Next.js or any framework
  • Zero runtime dependencies; ships as @mfukala/wsfs/client and @mfukala/wsfs/server

Sample use cases

  • Offline-friendly dashboards or CRMs that sync JSON blobs to an API while users travel
    Queries: "browser offline sync indexeddb rest etag", "local first crm cache optimistic concurrency", "web app sync json files"
  • Collaborative notes/tasks where conflicts should surface but not auto-merge
    Queries: "conflict aware offline note app etag", "optimistic concurrency file sync web", "custom conflict event listener indexeddb"
  • Self-hosted product data/settings that must work in service workers and fall back when the backend is down
    Queries: "service worker offline filesystem", "local first config storage web", "sync files nextjs api route if-match"
  • Binary or encrypted payloads that need custom codecs before storage/transfer
    Queries: "encrypt before indexeddb storage", "compress api payload sync", "browser base64 file sync"
  • Rapid prototyping of syncable file trees without picking a database upfront
    Queries: "virtual filesystem over http", "simple rest file sync adapter", "express etag file api starter"

The package ships:

  • A browser client (@mfukala/wsfs/client) that caches reads/writes locally, syncs in batches, and surfaces conflicts
  • A server toolkit (@mfukala/wsfs/server) that wires the sync protocol into any persistence adapter (an in-memory adapter is included)
  • No runtime dependencies

Browser client (@mfukala/wsfs/client)

import { Wsfs } from "@mfukala/wsfs/client";

const wsfs = await Wsfs.init({
  namespace: "vault", // IndexedDB namespace
  backendUrl: "http://localhost:8787", // your API base URL
});

// Mutations run inside a write task (exclusive lock + rollback on task failure).
await wsfs.runWriteTask(async (fs) => {
  await fs.write("/primary/item.json", JSON.stringify({ hello: "world" }));
});

// Reads run in read tasks (can run concurrently).
const item = await wsfs.runReadTask((fs) => fs.read("/primary/item.json"));

// Batch push local changes and pull remote updates.
await wsfs.sync();
  • runWriteTaskAndSync keeps the write lock through the sync; local edits rollback only if the task throws, not if sync fails.
  • list(prefix?) hides local tombstones; info(path) returns { etag, encoding, updatedBy? }. Use readMany(paths) / infoMany(paths) inside read tasks to fetch multiple files/metadata in one round-trip and cache the results—missing entries resolve to null instead of throwing.
  • Conflicts surface via a conflict event:
wsfs.addEventListener("conflict", (event) => {
  console.warn("Conflict detected", event.detail);
});
  • Pass a codec (encode/decode) to encrypt/compress payloads before hitting storage or the network.
  • Use attachAuth(kind, payload) to inject auth headers or signed proof fields on each request; pair it with the server-side authorize hook.

Attach auth headers/proofs before each request:

const signer = (input: { content: string | undefined; ifMatch?: string }) =>
  sign(JSON.stringify(input)); // your HMAC/EdDSA signer

const wsfs = await Wsfs.init({
  namespace: "vault",
  backendUrl: "http://localhost:8787",
  attachAuth: (kind, payload) => {
    payload.headers.Authorization = `Bearer ${accessToken}`;
    if (kind === "sync") {
      payload.body.writes?.forEach((write) => {
        write.proof = signer({ content: write.content ?? write.contentBase64, ifMatch: write.ifMatch });
      });
      payload.body.deletes?.forEach((del) => {
        del.proof = signer({ content: undefined, ifMatch: del.ifMatch });
      });
    }
  },
});

Server toolkit (@mfukala/wsfs/server)

Hook the protocol into any Node.js framework. createWsfsApi exposes the core sync/read/write/delete/list methods and works with any PersistenceAdapter. An in-memory adapter ships with the package for testing.

Express server example

import express from "express";
import { createWsfsApi, MemoryPersistence } from "@mfukala/wsfs/server";

const persistence = new MemoryPersistence(); // bring your own adapter in production
const api = createWsfsApi(persistence, {
  authorize: (kind, { headers, body }) => {
    // Verify auth headers/proofs before touching persistence.
    if ((headers?.["authorization"] ?? headers?.["Authorization"]) !== "Bearer secret") {
      const err = new Error("Unauthorized");
      (err as Error & { status?: number }).status = 401;
      throw err;
    }
    if (kind === "sync" && body && "writes" in body) {
      // Example: verify signatures attached by the client (body.writes[].proof)
    }
  },
  partition: ({ headers }) => {
    const tenant = headers?.["x-tenant"];
    return tenant ? { namespace: Array.isArray(tenant) ? tenant[0] : tenant } : undefined;
  },
});
const app = express();
app.use(express.json({ limit: "5mb" }));

app.post("/sync", async (req, res) => {
  try {
    res.status(200).json(await api.sync(req.body, { headers: req.headers }));
  } catch (err: any) {
    res.status(err?.status ?? 500).json({ error: err?.message ?? "sync failed" });
  }
});

app.get("/file", async (req, res) => {
  const file = await api.getFile(String(req.query.path ?? ""), { headers: req.headers });
  if (!file) return res.status(404).end();
  res.json(file);
});

app.get("/file/info", async (req, res) => {
  const info = await api.getFileInfo(String(req.query.path ?? ""), { headers: req.headers });
  if (!info) return res.status(404).end();
  res.json(info);
});

app.post("/file/batch", async (req, res) => {
  try {
    const paths = Array.isArray(req.body?.paths) ? req.body.paths : [];
    res.json(await api.getFiles(paths, { headers: req.headers }));
  } catch (err: any) {
    res.status(err?.status ?? 500).json({ error: err?.message ?? "getFiles failed" });
  }
});

app.post("/file/info/batch", async (req, res) => {
  try {
    const paths = Array.isArray(req.body?.paths) ? req.body.paths : [];
    res.json(await api.getFileInfos(paths, { headers: req.headers }));
  } catch (err: any) {
    res.status(err?.status ?? 500).json({ error: err?.message ?? "getFileInfos failed" });
  }
});

app.get("/list", async (req, res) => {
  res.json(await api.list(String(req.query.prefix ?? "/"), { headers: req.headers }));
});

app.put("/file", async (req, res) => {
  try {
    const result = await api.putFile(
      {
        path: req.body.path,
        content: req.body.content,
        contentBase64: req.body.contentBase64,
        encoding: req.body.encoding,
        ifMatch: req.headers["if-match"] as string | undefined,
        updatedBy: req.body.updatedBy,
        proof: req.body.proof, // arbitrary extra fields are allowed
      },
      { headers: req.headers },
    );
    res.status(200).json(result);
  } catch (err: any) {
    res.status(err?.status ?? 500).json({ error: err?.message ?? "put failed" });
  }
});

app.delete("/file", async (req, res) => {
  try {
    await api.deleteFile(
      {
        path: String(req.query.path ?? ""),
        ifMatch: req.headers["if-match"] as string | undefined,
        proof: req.body?.proof,
      },
      { headers: req.headers },
    );
    res.status(204).end();
  } catch (err: any) {
    res.status(err?.status ?? 500).json({ error: err?.message ?? "delete failed" });
  }
});

app.listen(8787);

Next.js API Route example

// pages/api/wsfs/sync.ts (or app/api/wsfs/sync/route.ts)
import { createWsfsApi, MemoryPersistence } from "@mfukala/wsfs/server";

const api = createWsfsApi(new MemoryPersistence()); // swap in your adapter

export default async function handler(req, res) {
  try {
    const result = await api.sync(req.body, { headers: req.headers });
    res.status(200).json(result);
  } catch (err: any) {
    res.status(err?.status ?? 500).json({ error: err?.message ?? "sync failed" });
  }
}

Reuse the same pattern for /file, /file/info, /file/batch, /file/info/batch, and /list, passing the request payloads into api.putFile, api.getFile, api.getFileInfo, api.getFiles, api.getFileInfos, and api.list while preserving the If-Match header for writes/deletes. authorize(kind, payload) runs before persistence and may throw with status (401/403/400) to block the request; partition(ctx) can select a tenant and falls back to the adapter’s baked-in partition when undefined.

Example signature check inside authorize:

const api = createWsfsApi(persistence, {
  authorize: (kind, { body }) => {
    if (kind !== "sync") return;
    for (const write of body.writes ?? []) {
      verifyProof(write.proof, {
        content: write.content ?? write.contentBase64,
        ifMatch: write.ifMatch,
      });
    }
    for (const del of body.deletes ?? []) {
      verifyProof(del.proof, { content: undefined, ifMatch: del.ifMatch });
    }
  },
});

Server endpoints the client expects

  • POST /sync with { prefix, writes, deletes, known, watermark }
  • GET /file?path=/foo.txt{ etag, encoding, updatedBy?, content|contentBase64 }
  • GET /file/info?path=/foo.txt{ etag, encoding, updatedBy? }
  • POST /file/batch + body { paths: ["/foo.txt", "/bar.txt"] }[EncodedRecord | null, ...]
  • POST /file/info/batch + body { paths: ["/foo.txt"] }[{ etag, encoding, updatedBy? } | null, ...]
  • GET /list?prefix=/dir/[{ path, etag, encoding }]
  • PUT /file + If-Match: <etag|*>{ etag }
  • DELETE /file?path=... + If-Match: <etag|*>

The bundled MemoryPersistence enforces If-Match (use "*" to create new files), tracks updatedBy, and exposes incremental sync via watermarks.

SQL persistence (driver-agnostic)

SqlPersistence stays runtime-agnostic: you supply a tiny executor (get/all/run and optional transaction) plus partition binders (e.g., user_id, vault_id). Bring your own driver (SQLite/Postgres/MySQL, etc.) and wire it up:

import Database from "better-sqlite3";
import { SqlPersistence, createWsfsApi } from "@mfukala/wsfs/server";

const db = new Database(":memory:");
// create `files` + `file_changes` tables with your partition columns

const persistence = new SqlPersistence({
  table: "files",
  changesTable: "file_changes",
  executor: {
    get: (sql, params) => db.prepare(sql).get(params),
    all: (sql, params) => db.prepare(sql).all(params),
    run: (sql, params) => {
      const { changes } = db.prepare(sql).run(params);
      return { rowsAffected: changes };
    },
    transaction: async (fn) => {
      db.exec("BEGIN");
      try {
        const result = await fn();
        db.exec("COMMIT");
        return result;
      } catch (err) {
        db.exec("ROLLBACK");
        throw err;
      }
    },
  },
  partition: { columns: ["user_id", "vault_id"], toParams: (p) => [p.userId, p.vaultId] },
  partitionValue: { userId: "demo", vaultId: "vault-1" },
});

const api = createWsfsApi(persistence);

For multi-tenant setups, keep a root adapter and let createWsfsApi call persistence.withPartition(...) via the partition hook (or call it yourself before wiring the API). Watermarks come from file_changes.updated_at, enabling incremental listChanges.

Transactions are optional; omit them with drivers that require synchronous callbacks (e.g., better-sqlite3) and rely on single-statement atomicity, or supply an async-friendly transaction wrapper when your driver supports it.