npm package discovery and stats viewer.

Discover Tips

  • General search

    [free text search, go nuts!]

  • Package details

    pkg:[package-name]

  • User packages

    @[username]

Sponsor

Optimize Toolset

I’ve always been into building performant and accessible sites, but lately I’ve been taking it extremely seriously. So much so that I’ve been building a tool to help me optimize and monitor the sites that I build to make sure that I’m making an attempt to offer the best experience to those who visit them. If you’re into performant, accessible and SEO friendly sites, you might like it too! You can check it out at Optimize Toolset.

About

Hi, 👋, I’m Ryan Hefner  and I built this site for me, and you! The goal of this site was to provide an easy way for me to check the stats on my npm packages, both for prioritizing issues and updates, and to give me a little kick in the pants to keep up on stuff.

As I was building it, I realized that I was actually using the tool to build the tool, and figured I might as well put this out there and hopefully others will find it to be a fast and useful way to search and browse npm packages as I have.

If you’re interested in other things I’m working on, follow me on Twitter or check out the open source projects I’ve been publishing on GitHub.

I am also working on a Twitter bot for this site to tweet the most popular, newest, random packages from npm. Please follow that account now and it will start sending out packages soon–ish.

Open Software & Tools

This site wouldn’t be possible without the immense generosity and tireless efforts from the people who make contributions to the world and share their work via open source initiatives. Thank you 🙏

© 2026 – Pkg Stats / Ryan Hefner

chimera-sigil

v0.1.3

Published

Multi-model AI agent harness for the terminal

Readme

Chimera Sigil

Chimera Sigil gives you one terminal workflow that starts with local Ollama models and can still reach Grok, OpenAI, or Anthropic when you explicitly choose them. It is built for short feedback loops, visible tool use, and fast experimentation without requiring API spend.

Mission

Ship a practical local-first multi-model harness that lets people get real work done for free on the hardware they already have, then scales up to bigger local GPUs or paid APIs only when they choose that path.

What It Does

  • Runs a terminal agent with built-in file, search, edit, and shell tools
  • Supports local model profiles, collaborator models, and fallback chains
  • Persists sessions so single-prompt runs can resume later
  • Supports approval modes for safer noninteractive automation
  • Builds release artifacts locally on macOS, Windows, and Linux/WSL

Quick Start

Install from npm:

npm install -g chimera-sigil
chimera --local-doctor
ollama pull qwen3:4b
chimera --help

Install and start Ollama first if it is not already running.

Run from source:

cargo run -p chimera-sigil-cli

Single prompt mode:

cargo run -p chimera-sigil-cli -- --prompt "Summarize this codebase"

Local coding profile:

cargo run -p chimera-sigil-cli -- --model local-coder --prompt "Find the risky parts of this patch"

Collaborative local mode:

cargo run -p chimera-sigil-cli -- --model local-small --collab local-tiny,local-coder

Fallback mode:

cargo run -p chimera-sigil-cli -- --model local-coder --fallback local-small,local-tiny

Local Model Profiles

The default CLI model is local, which resolves to qwen3:4b through Ollama. These aliases are intentionally conservative so a person with a newer budget laptop can start useful work without buying API credits.

| Profile | Ollama model | Best fit | | --- | --- | --- | | local-tiny | llama3.2:1b | Lowest memory, quick notes, routing, simple edits | | local-edge | gemma3n:e2b | Everyday-device profile for laptops and small machines | | local, local-small, local-laptop | qwen3:4b | Default free laptop profile | | local-coder-small | qwen2.5-coder:3b | Lightweight coding assistant | | local-coder, local-code | qwen2.5-coder:7b | Better code review and edits on 16GB+ machines | | local-balanced | qwen3:8b | Stronger general work when memory allows | | local-12gb, local-16gb, local-heavy | qwen3:14b | Midrange NVIDIA GPUs and larger-memory laptops | | local-coder-12gb, local-coder-16gb, local-coder-heavy | qwen2.5-coder:14b | Midrange GPU coding profile | | local-reasoning | deepseek-r1:8b | Slower reasoning pass for hard problems | | local-24gb, local-4090, local-workstation, local-gpu | qwen3:30b | 24GB GPU generalist profile | | local-coder-24gb, local-coder-4090, local-coder-gpu | qwen2.5-coder:32b | 24GB GPU coding profile |

Any Ollama model name also works directly:

chimera --model gemma3n
chimera --model qwen3:8b
chimera --model qwen2.5-coder:14b

For the full hardware matrix and benchmarking plan, see docs/hardware-optimization-process.md.

Ask Chimera to recommend a local profile for the current machine:

chimera --local-doctor
chimera --local-doctor --json

Benchmark local profiles with Ollama timing metrics:

chimera --local-benchmark
chimera --local-benchmark --benchmark-models local-4090,local-coder-4090,local-coder-16gb

Approval Modes

  • --approval-mode prompt: ask before writes and command execution
  • --approval-mode approve: allow workspace writes, deny shell execution
  • --approval-mode full: allow all tools
  • --auto-approve: compatibility shorthand for --approval-mode full

Environment

Local Ollama needs no API key. Cloud providers are optional:

export OLLAMA_BASE_URL=http://localhost:11434/v1
export XAI_API_KEY=...
export OPENAI_API_KEY=...
export ANTHROPIC_API_KEY=...

No provider keys are committed to this repository. Local machine, Ollama, and release credentials live outside the repo.

Fast Iteration

  • Prompt mode auto-saves sessions for later resume
  • Release builds can be produced locally instead of relying on GitHub Actions
  • The shipped CLI command stays short: chimera

Contributing

Contributions are welcome. Keep changes small, tested, and easy to review.

Start here:

Project Layout

  • crates/chimera-core: agent loop, config, sessions
  • crates/chimera-providers: model providers and streaming
  • crates/chimera-tools: built-in tools and permissions
  • crates/chimera-cli: CLI entrypoint
  • scripts/: packaging and local-first release tooling

Release Notes

The npm package name is chimera-sigil. The installed command is chimera.

Local-first release tooling currently supports:

  • aarch64-apple-darwin
  • x86_64-apple-darwin
  • x86_64-pc-windows-msvc
  • x86_64-unknown-linux-gnu

For contributor setup and release workflow details, see CONTRIBUTING.md.

Support

Everything here is released for public use. If Chimera saved you time or you want to keep the work moving, you can support public FRG releases.