npm package discovery and stats viewer.

Discover Tips

  • General search

    [free text search, go nuts!]

  • Package details

    pkg:[package-name]

  • User packages

    @[username]

Sponsor

Optimize Toolset

I’ve always been into building performant and accessible sites, but lately I’ve been taking it extremely seriously. So much so that I’ve been building a tool to help me optimize and monitor the sites that I build to make sure that I’m making an attempt to offer the best experience to those who visit them. If you’re into performant, accessible and SEO friendly sites, you might like it too! You can check it out at Optimize Toolset.

About

Hi, 👋, I’m Ryan Hefner  and I built this site for me, and you! The goal of this site was to provide an easy way for me to check the stats on my npm packages, both for prioritizing issues and updates, and to give me a little kick in the pants to keep up on stuff.

As I was building it, I realized that I was actually using the tool to build the tool, and figured I might as well put this out there and hopefully others will find it to be a fast and useful way to search and browse npm packages as I have.

If you’re interested in other things I’m working on, follow me on Twitter or check out the open source projects I’ve been publishing on GitHub.

I am also working on a Twitter bot for this site to tweet the most popular, newest, random packages from npm. Please follow that account now and it will start sending out packages soon–ish.

Open Software & Tools

This site wouldn’t be possible without the immense generosity and tireless efforts from the people who make contributions to the world and share their work via open source initiatives. Thank you 🙏

© 2026 – Pkg Stats / Ryan Hefner

codemaxxing

v1.1.1

Published

Open-source terminal coding agent. Connect any LLM. Max your code.

Readme

codemaxxing 💪

your code. your model. no excuses.

npm version license

Open-source terminal coding agent. Connect any LLM — local or remote — and start building. Like Claude Code, but you bring your own model.

🆕 v1.1.0: Use GPT-5.4 with your ChatGPT Plus subscription — no API key needed. Just /login → OpenAI → OAuth. Same access as Codex CLI.

Why?

Every coding agent locks you into their API. Codemaxxing doesn't. Run it with LM Studio, Ollama, OpenRouter, OpenAI, Anthropic, or any OpenAI-compatible endpoint. Your machine, your model, your rules.

Install

If you have Node.js:

npm install -g codemaxxing

If you don't have Node.js:

The one-line installers below will install Node.js first, then codemaxxing.

Linux / macOS:

bash -c "$(curl -fsSL https://raw.githubusercontent.com/MarcosV6/codemaxxing/main/install.sh)"

Windows (CMD as Administrator):

curl -fsSL -o %TEMP%\install-codemaxxing.bat https://raw.githubusercontent.com/MarcosV6/codemaxxing/main/install.bat && %TEMP%\install-codemaxxing.bat

Windows (PowerShell as Administrator):

curl -fsSL -o $env:TEMP\install-codemaxxing.bat https://raw.githubusercontent.com/MarcosV6/codemaxxing/main/install.bat; & $env:TEMP\install-codemaxxing.bat

Windows note: If Node.js was just installed, you may need to close and reopen your terminal, then run npm install -g codemaxxing manually. This is a Windows PATH limitation.

Updating

npm update -g codemaxxing

If that doesn't get the latest version, use the exact reinstall path:

npm install -g codemaxxing@latest

Then verify:

codemaxxing --version

Quick Start

Option A — easiest local setup

If you already have a local server running, Codemaxxing auto-detects common defaults:

  • LM Studio on http://localhost:1234/v1
  • Ollama on http://localhost:11434
  • vLLM on http://localhost:8000

For LM Studio:

  1. Download LM Studio
  2. Load a coding model (for example Qwen 2.5 Coder 7B for a lightweight test)
  3. Start the local server
  4. Run:
codemaxxing

Option B — no local model yet

Just run:

codemaxxing

If no LLM is available, Codemaxxing can guide you through:

  • detecting your hardware
  • recommending a model
  • installing Ollama
  • downloading the model
  • connecting automatically

Option C — ChatGPT Plus (GPT-5.4, easiest cloud option)

If you have a ChatGPT Plus subscription, get instant access to GPT-5.4 with zero API costs:

codemaxxing login
# → Pick "OpenAI"
# → Pick "OpenAI (ChatGPT)" 
# → Browser opens, log in with your ChatGPT account
# → Done — you now have GPT-5.4, GPT-5, o3, o4-mini

codemaxxing
# → /model → pick gpt-5.4

No API key required. Uses your ChatGPT subscription limits instead.

Option D — other cloud providers

Authenticate first:

codemaxxing login

Then run:

codemaxxing

Authentication

One command to connect any provider:

codemaxxing login

Interactive setup walks you through it. Or use /login inside the TUI.

Supported auth methods:

| Provider | Methods | |----------|---------| | OpenRouter | OAuth (browser login) or API key — one login, 200+ models | | Anthropic | Link your Claude subscription (via Claude Code) or API key | | OpenAI | Import from Codex CLI or API key | | Qwen | Import from Qwen CLI or API key | | GitHub Copilot | Device flow (browser) | | Google Gemini | API key | | Any provider | API key + custom base URL |

codemaxxing login              # Interactive provider picker
codemaxxing auth list          # See saved credentials
codemaxxing auth remove <name> # Delete a credential
codemaxxing auth openrouter    # Direct OpenRouter OAuth

Credentials stored securely in ~/.codemaxxing/auth.json (owner-only permissions).


Advanced Setup

With a remote provider (OpenAI, OpenRouter, etc.):

codemaxxing --base-url https://api.openai.com/v1 --api-key sk-... --model gpt-5

With a saved provider profile:

codemaxxing --provider openrouter

Auto-detected local servers: LM Studio (:1234), Ollama (:11434), vLLM (:8000)

Features

🔥 Streaming Tokens

Real-time token display. See the model think, not just the final answer.

⚠️ Tool Approval + Diff Preview

Dangerous operations require your approval. File writes show a unified diff of what will change before you say yes. Press y to allow, n to deny, a to always allow.

🏗️ Architect Mode

Dual-model planning. A "planner" model reasons through the approach, then your editor model executes the changes.

  • /architect — toggle on/off
  • /architect claude-3-5-sonnet — set the planner model
  • Great for pairing expensive reasoning models with fast editors

🧠 Skills System (21 Built-In)

Downloadable skill packs that teach the agent domain expertise. Ships with 21 built-in skills and a menu-first /skills flow so you can browse instead of memorizing names:

| Category | Skills | |----------|--------| | Frontend | react-expert, nextjs-app, tailwind-ui, svelte-kit | | Mobile | react-native, swift-ios, flutter | | Backend | python-pro, node-backend, go-backend, rust-systems | | Data | sql-master, supabase | | Practices | typescript-strict, api-designer, test-engineer, doc-writer, security-audit, devops-toolkit, git-workflow | | Game Dev | unity-csharp |

/skills              # Browse & install from registry
/skills install X    # Quick install
/skills on/off X     # Toggle per session

📋 CODEMAXXING.md — Project Rules

Drop a CODEMAXXING.md in your project root for project-specific instructions. It gets loaded automatically for that project.

🔧 Auto-Lint

Automatically runs your linter after every file edit and feeds errors back to the model for auto-fix. Detects eslint, biome, ruff, clippy, golangci-lint, and more.

  • /lint on / /lint off — toggle (ON by default)

📂 Smart Context (Repo Map)

Scans your codebase and builds a map of functions, classes, and types. The model knows what exists where without reading every file.

📦 Context Compression

When conversation history gets too large, older messages are automatically summarized to free up context.

💰 Cost Tracking

Per-session token usage and estimated cost in the status bar. Pricing for 20+ common models. Saved to session history.

🖥️ Headless/CI Mode

Run codemaxxing in scripts and pipelines without the TUI:

codemaxxing exec "add error handling to api.ts"
codemaxxing exec --auto-approve "fix all lint errors"
echo "add tests" | codemaxxing exec

🔀 Git Integration

Opt-in git commands built in:

  • /commit <message> — stage all + commit
  • /push — push to remote
  • /diff — show changes
  • /undo — revert last codemaxxing commit
  • /git on / /git off — toggle auto-commits

💾 Session Persistence

Conversations auto-save to SQLite. Pick up where you left off:

  • /sessions — list past sessions
  • /session delete — remove a session
  • /resume — interactive session picker

🔌 MCP Support (Model Context Protocol)

Connect to external tools via the MCP standard: databases, GitHub, Slack, browsers, and more.

  • /mcp — show connected servers
  • /mcp add github npx -y @modelcontextprotocol/server-github — add a server
  • /mcp tools — list available MCP tools

🖥️ Zero-Setup Local LLM

First time with no LLM? Codemaxxing walks you through it:

  1. Detects your hardware (CPU, RAM, GPU)
  2. Recommends coding models that fit your machine
  3. Installs Ollama automatically
  4. Downloads the model with a progress bar
  5. Connects and drops you into coding mode

No googling, no config files, no decisions. Just run codemaxxing.

🦙 Ollama Management

Full Ollama control from inside codemaxxing:

  • /ollama — status, installed models, GPU usage
  • /ollama pull — interactive model picker + download
  • /ollama delete — pick and remove models
  • /ollama start / /ollama stop — server management
  • Exit warning when Ollama is using GPU memory

🔄 Multi-Provider

Switch models mid-session with an interactive picker:

  • /model — browse and switch models
  • /model gpt-5 — switch directly by name
  • Native Anthropic API support (not just OpenAI-compatible)

🎨 14 Themes

/theme to browse: cyberpunk-neon, dracula, gruvbox, nord, catppuccin, tokyo-night, one-dark, rose-pine, synthwave, blood-moon, mono, solarized, hacker, acid

🔐 Authentication

One command to connect any LLM provider. OpenRouter OAuth, Anthropic subscription linking, Codex/Qwen CLI import, GitHub Copilot device flow, or manual API keys.

📋 Smart Paste

Multi-line pastes collapse into [Pasted text #1 +N lines] badges instead of dumping raw text into the input box. This was specifically hardened for bracketed-paste terminal weirdness.

⌨️ Slash Commands

Type / for autocomplete suggestions. Arrow keys to navigate, Tab or Enter to select.

Commands

| Command | Description | |---------|-------------| | /help | Show all commands | | /connect | Retry LLM connection | | /login | Interactive auth setup | | /model | Browse & switch models (picker) | | /architect | Toggle architect mode / set model | | /skills | Browse, install, manage skills | | /lint on/off | Toggle auto-linting | | /mcp | MCP server status & tools | | /ollama | Ollama status, models & GPU | | /ollama pull | Download a model (picker) | | /ollama delete | Remove a model (picker) | | /ollama start/stop | Server management | | /theme | Switch color theme | | /map | Show repository map | | /sessions | List past sessions | | /session delete | Delete a session | | /resume | Resume a past session | | /reset | Clear conversation | | /context | Show message count + tokens | | /diff | Show git changes | | /commit <msg> | Stage all + commit | | /push | Push to remote | | /undo | Revert last codemaxxing commit | | /git on/off | Toggle auto-commits | | /quit | Exit |

CLI

codemaxxing                          # Start TUI
codemaxxing login                    # Auth setup
codemaxxing auth list                # Show saved credentials
codemaxxing exec "prompt"            # Headless mode (no TUI)
codemaxxing exec --auto-approve "x"  # Skip approval prompts
codemaxxing exec --json "x"          # JSON output for scripts
echo "fix tests" | codemaxxing exec  # Pipe from stdin
-m, --model <model>       Model name to use
-p, --provider <name>     Provider profile from config
-k, --api-key <key>       API key for the provider
-u, --base-url <url>      Base URL for the provider API
-h, --help                Show help

Config

Settings are stored in ~/.codemaxxing/settings.json:

{
  "provider": {
    "baseUrl": "http://localhost:1234/v1",
    "apiKey": "not-needed",
    "model": "auto"
  },
  "providers": {
    "local": {
      "name": "Local (LM Studio/Ollama)",
      "baseUrl": "http://localhost:1234/v1",
      "apiKey": "not-needed",
      "model": "auto"
    },
    "openrouter": {
      "name": "OpenRouter",
      "baseUrl": "https://openrouter.ai/api/v1",
      "apiKey": "sk-or-...",
      "model": "anthropic/claude-sonnet-4-6"
    },
    "openai": {
      "name": "OpenAI",
      "baseUrl": "https://api.openai.com/v1",
      "apiKey": "sk-...",
      "model": "gpt-5"
    }
  },
  "defaults": {
    "autoApprove": false,
    "maxTokens": 8192
  }
}

Tools

Built-in tools:

  • read_file — Read file contents (safe)
  • write_file — Write/create files (requires approval, shows diff)
  • edit_file — Apply surgical patches to files (preferred for targeted changes)
  • list_files — List directory contents (safe)
  • search_files — Search for patterns across files (safe)
  • run_command — Execute shell commands (requires approval)

Plus any tools from connected MCP servers (databases, APIs, GitHub, etc.)

Project Context

Drop a CODEMAXXING.md file in your project root to give the model extra context about your codebase, conventions, or instructions. It's automatically included in the system prompt.

Stack

Inspired By

Built by studying the best:

  • Aider — repo map concept, auto-commit
  • Claude Code — permission system, paste handling
  • OpenCode — multi-provider, SQLite sessions

License

MIT