npm package discovery and stats viewer.

Discover Tips

  • General search

    [free text search, go nuts!]

  • Package details

    pkg:[package-name]

  • User packages

    @[username]

Sponsor

Optimize Toolset

I’ve always been into building performant and accessible sites, but lately I’ve been taking it extremely seriously. So much so that I’ve been building a tool to help me optimize and monitor the sites that I build to make sure that I’m making an attempt to offer the best experience to those who visit them. If you’re into performant, accessible and SEO friendly sites, you might like it too! You can check it out at Optimize Toolset.

About

Hi, 👋, I’m Ryan Hefner  and I built this site for me, and you! The goal of this site was to provide an easy way for me to check the stats on my npm packages, both for prioritizing issues and updates, and to give me a little kick in the pants to keep up on stuff.

As I was building it, I realized that I was actually using the tool to build the tool, and figured I might as well put this out there and hopefully others will find it to be a fast and useful way to search and browse npm packages as I have.

If you’re interested in other things I’m working on, follow me on Twitter or check out the open source projects I’ve been publishing on GitHub.

I am also working on a Twitter bot for this site to tweet the most popular, newest, random packages from npm. Please follow that account now and it will start sending out packages soon–ish.

Open Software & Tools

This site wouldn’t be possible without the immense generosity and tireless efforts from the people who make contributions to the world and share their work via open source initiatives. Thank you 🙏

© 2026 – Pkg Stats / Ryan Hefner

moltmind

v0.8.3

Published

Agent Memory MCP Server — persistent semantic memory and session continuity for AI agents

Readme

MoltMind

Persistent semantic memory and session continuity for AI agents. One install, zero config, runs 100% locally.

MoltMind is an MCP server that gives your AI agent long-term memory across sessions — storing learnings, decisions, error fixes, and handoff context using local SQLite and embeddings. No API keys, no cloud, no accounts needed.

Why MoltMind?

Every time your AI agent starts a new conversation, it forgets everything. It spends 1-2 minutes re-reading your files, re-learning your architecture, and re-discovering decisions you already made. MoltMind gives it memory — your agent picks up right where it left off in seconds.

| | Without MoltMind | With MoltMind | |--|-----------------|---------------| | Model used | Claude Opus 4.6 ($5/$25 per 1M tokens) | | | Time per session | 1-2 min re-exploring | Seconds to resume | | Cost per session | ~$0.09 | ~$0.009 | | 20 sessions | $1.80 | $0.18 | | Daily use (1 year) | $32.85 | $3.29 | | Time saved (1 year) | — | ~6 hours | | Money saved (1 year) | — | ~$30 |

Assumes ~8,000 input + ~2,000 output tokens per cold start, ~825 input + ~200 output per resume. Savings scale with usage — power users save more.

Quick Start

Claude Code

claude mcp add moltmind -- npx -y moltmind

Restart Claude Code, then run /mcp to verify.

With moltbook social features:

claude mcp add moltmind -- npx -y moltmind --moltbook

See moltbook.com for the agent social network.

Other Clients

Add to your client's MCP config file:

{
  "mcpServers": {
    "moltmind": {
      "command": "npx",
      "args": ["-y", "moltmind"]
    }
  }
}

With moltbook: "args": ["-y", "moltmind", "--moltbook"]

| Client | Config file | Key | |--------|------------|-----| | Cursor | ~/.cursor/mcp.json or .cursor/mcp.json | mcpServers | | Windsurf | ~/.codeium/windsurf/mcp_config.json | mcpServers | | VS Code (Copilot) | .vscode/mcp.json | servers | | Cline | Settings > MCP Servers > Edit Config | mcpServers | | Codex CLI | ~/.codex/config.json | mcpServers | | Any MCP client | Varies | mcpServers |

Note: VS Code uses "servers" instead of "mcpServers" as the top-level key.

Upgrading

npx -y moltmind always fetches the latest version. If you have a stale global install:

npm uninstall -g moltmind   # then let npx handle it

Tools

14 core tools by default, 21 with --moltbook:

| Tool | Description | |------|-------------| | mm_store | Store a memory (learning, error fix, decision, plan, or raw note) | | mm_recall | Search memories — hybrid semantic + keyword search | | mm_read | Read a specific memory by ID | | mm_update | Update an existing memory | | mm_delete | Archive a memory (soft delete) | | mm_status | Server health dashboard | | mm_init | Create a project-local memory vault | | mm_handoff_create | Structured handoff for agent-to-agent context transfer | | mm_handoff_load | Load the most recent handoff | | mm_session_save | Save session summary and where you left off | | mm_session_resume | Restore context from recent sessions | | mm_session_history | Browse past sessions with tool call stats | | mm_feedback | Report bugs or request features | | mm_metrics | Adoption and health metrics dashboard |

How It Works

Memory & Search — Your agent stores memories in a local database. When it needs to find something, MoltMind searches by meaning (not just keywords) — so searching for "API port" finds a memory about "our server runs on port 8080". If the search model isn't downloaded yet, it falls back to keyword matching.

Sessions & Handoffs — Sessions are auto-created on startup and auto-paused on shutdown. Your agent saves where it left off and picks up seamlessly next time. Handoffs let one agent pass context to another with structured goal/state/next-action documents.

Diagnostics — Every tool call is logged locally with timing and success/failure. mm_status shows health, mm_metrics shows usage stats and token savings. All data stays on your machine.

What It Costs (Tokens)

Every MCP tool adds a small overhead to each request because the AI needs to know what tools are available. Here's what MoltMind costs — and what it saves you:

| | Cost per request | In dollars | |--|-----------------|------------| | MoltMind overhead (14 tools) | ~500 tokens | ~$0.0015 | | With prompt caching | ~50 tokens | ~$0.00015 | | Session resume (saves you) | ~7,675 tokens | ~$0.023 |

Bottom line: MoltMind pays for itself after a single session resume. Every conversation after that is pure savings.

Free vs Pro

| | Free | Pro ($9 one-time) | |--|------|-----| | Stores per day | 20 | Unlimited | | Total memories | 200 | Unlimited | | Search | Unlimited | Unlimited | | Session tools | Unlimited | Unlimited | | Vector search | Brute-force | Zvec ANN (auto) |

Upgrade to Pro

From your terminal

macOS / Linux:

npx moltmind --upgrade

Windows (Command Prompt or PowerShell):

npx moltmind --upgrade

Same command on all platforms. A checkout page opens in your browser, you pay $9, and the terminal waits up to 5 minutes for payment. Once confirmed, your license activates automatically — no copy-pasting keys, no config files.

Prerequisite: You must have run MoltMind at least once before upgrading (so your machine ID exists at ~/.moltmind/instance_id). If you just installed, open your AI client, let MoltMind start, then come back and run --upgrade.

From inside your AI agent

If you're already in a conversation with an agent that has MoltMind installed, just say:

"Upgrade MoltMind to Pro"

Your agent will run npx moltmind --upgrade for you, the checkout page opens, and the license activates after payment. Works in Claude Code, Cursor, Windsurf, Cline, Codex CLI — any client where MoltMind is installed.

After payment

  • License is written to ~/.moltmind/license.key
  • Zvec ANN search auto-enables on next startup
  • mm_status shows "Pro" tier
  • Restart your MCP client to pick up the change

Machine lock and migration

Pro licenses are tied to one machine (your instance_id). A daily heartbeat check enforces this.

  • Switching machines? Run npx moltmind --upgrade on the new one. Your old machine reverts to free within 24 hours.
  • Offline? A 7-day grace period keeps Pro active while you're disconnected.
  • All data stays local. The heartbeat only sends your machine ID and license prefix — no memory content ever leaves your machine.

Search Performance (Pro)

Pro tier uses Zvec ANN for fast memory search. Here's what that means in practice:

Accuracy — At 1,000 memories (a typical heavy user), Zvec finds 98% of the exact same results as an exhaustive search. Your agent gets the right answer almost every time.

Speed — Search takes under 1ms at 1,000 memories. At 10,000 memories, it's still under 5ms. Your agent won't notice any delay.

Reliability — Handles 330+ searches per second with zero latency spikes. Deleted memories never come back. Results are deterministic.

See BENCHMARK_RESULTS.md for the full report, or RUNBOOK.md for how to run benchmarks yourself.

Data Storage

| Path | Purpose | |------|---------| | ~/.moltmind/memory.db | Global memory vault | | ./.moltmind/memory.db | Project-local vault (via mm_init) | | ~/.moltmind/models/ | Cached embedding model (~22MB) | | ~/.moltmind/instance_id | Anonymous instance identifier | | ~/.moltmind/license.key | Pro license (after upgrade) |

Requirements

  • Node.js 18+
  • No API keys required
  • No internet after first embedding model download

License

MIT