npm package discovery and stats viewer.

Discover Tips

  • General search

    [free text search, go nuts!]

  • Package details

    pkg:[package-name]

  • User packages

    @[username]

Sponsor

Optimize Toolset

I’ve always been into building performant and accessible sites, but lately I’ve been taking it extremely seriously. So much so that I’ve been building a tool to help me optimize and monitor the sites that I build to make sure that I’m making an attempt to offer the best experience to those who visit them. If you’re into performant, accessible and SEO friendly sites, you might like it too! You can check it out at Optimize Toolset.

About

Hi, 👋, I’m Ryan Hefner  and I built this site for me, and you! The goal of this site was to provide an easy way for me to check the stats on my npm packages, both for prioritizing issues and updates, and to give me a little kick in the pants to keep up on stuff.

As I was building it, I realized that I was actually using the tool to build the tool, and figured I might as well put this out there and hopefully others will find it to be a fast and useful way to search and browse npm packages as I have.

If you’re interested in other things I’m working on, follow me on Twitter or check out the open source projects I’ve been publishing on GitHub.

I am also working on a Twitter bot for this site to tweet the most popular, newest, random packages from npm. Please follow that account now and it will start sending out packages soon–ish.

Open Software & Tools

This site wouldn’t be possible without the immense generosity and tireless efforts from the people who make contributions to the world and share their work via open source initiatives. Thank you 🙏

© 2026 – Pkg Stats / Ryan Hefner

@letchu_pkt/claudex

v1.1.0

Published

Claude Code opened to any LLM — OpenAI, Gemini, DeepSeek, Ollama, and 200+ models

Readme

Claudex

Claudex is a fork of the Claude Code source that adds a full OpenAI-compatible provider shim, NVIDIA AI (NIM) support, a smart multi-provider router, a Telegram gateway, and local inference via Ollama and Atomic Chat. Every Claude Code tool works — bash, file ops, grep, glob, agents, MCP, tasks — powered by whatever model you choose.

Author: Lakshmikanthan K — github.com/l3tchupkt


Install

npm install -g @letchu_pkt/claudex
claudex

On first run, if no provider is configured, Claudex will prompt you to set one up via /provider. No Anthropic account required.


Fastest Start

Pick a provider, set three env vars, run.

NVIDIA AI — free key, best models

export CLAUDE_CODE_USE_NVIDIA=1
export NVIDIA_API_KEY=nvapi-your-key
export NVIDIA_MODEL=moonshotai/kimi-k2-instruct
claudex

Free key at build.nvidia.com.

OpenAI

export CLAUDE_CODE_USE_OPENAI=1
export OPENAI_API_KEY=sk-your-key
export OPENAI_MODEL=gpt-4o
claudex

Google Gemini — free key

export CLAUDE_CODE_USE_GEMINI=1
export GEMINI_API_KEY=your-key
export GEMINI_MODEL=gemini-2.0-flash
claudex

Free key at aistudio.google.com/apikey.

Ollama — fully local, no key

ollama pull llama3.1:8b
export CLAUDE_CODE_USE_OPENAI=1
export OPENAI_BASE_URL=http://localhost:11434/v1
export OPENAI_MODEL=llama3.1:8b
claudex

Guides

| Audience | Guide | |---|---| | New to terminals | Non-Technical Setup | | Windows | Windows Quick Start | | macOS / Linux | macOS / Linux Quick Start | | Android (Termux) | Android Install | | Source builds, profiles, diagnostics | Advanced Setup | | Daily Ollama workflow | Playbook | | Telegram bot | Telegram Gateway |


Supported Providers

| Provider | Env flag | Key var | |---|---|---| | NVIDIA AI (NIM) | CLAUDE_CODE_USE_NVIDIA=1 | NVIDIA_API_KEY | | OpenAI / any OpenAI-compatible | CLAUDE_CODE_USE_OPENAI=1 | OPENAI_API_KEY | | Google Gemini | CLAUDE_CODE_USE_GEMINI=1 | GEMINI_API_KEY | | GitHub Models | CLAUDE_CODE_USE_GITHUB=1 | GITHUB_TOKEN | | Amazon Bedrock | CLAUDE_CODE_USE_BEDROCK=1 | AWS credentials | | Google Vertex AI | CLAUDE_CODE_USE_VERTEX=1 | GCP credentials | | Microsoft Foundry | CLAUDE_CODE_USE_FOUNDRY=1 | ANTHROPIC_FOUNDRY_API_KEY | | Ollama (local) | CLAUDE_CODE_USE_OPENAI=1 + localhost URL | none | | Atomic Chat (Apple Silicon) | CLAUDE_CODE_USE_OPENAI=1 + 127.0.0.1:1337 | none | | Anthropic (default) | none | ANTHROPIC_API_KEY |

Any OpenAI-compatible endpoint works: DeepSeek, Groq, Mistral, Together AI, OpenRouter, LM Studio, Azure OpenAI, and more.


NVIDIA AI Models

| Model | Best for | |---|---| | moonshotai/kimi-k2-instruct | Reasoning, coding (default) | | nvidia/llama-3.1-nemotron-ultra-253b-v1 | Flagship quality | | meta/llama-3.3-70b-instruct | Balanced speed/quality | | meta/llama-3.1-8b-instruct | Fast, lightweight | | deepseek-ai/deepseek-r1 | Deep reasoning | | qwen/qwen3-235b-a22b | Large MoE | | mistralai/mistral-large-2-instruct | Instruction following |


Startup Themes

Set CLAUDEX_THEME to change the banner color scheme:

| Theme | Colors | |---|---| | sunset | warm orange → rust (default) | | ocean | deep teal → electric cyan | | aurora | green → violet | | neon | hot pink → electric blue | | mono | white → grey |

export CLAUDEX_THEME=ocean
claudex

Profile Launcher

Save a provider profile once, launch with one command:

# save a profile (also works via /provider inside the CLI)
bun run profile:init -- --provider nvidia --api-key nvapi-...
bun run profile:init -- --provider openai --api-key sk-...
bun run profile:init -- --provider ollama --model llama3.1:8b

# launch from saved profile
bun run dev:profile

# provider-specific launchers
bun run dev:nvidia
bun run dev:openai
bun run dev:ollama
bun run dev:gemini
bun run dev:codex
bun run dev:atomic-chat

Smart Router

Benchmarks all configured providers on startup and routes each request to the fastest, cheapest, healthiest option:

export ROUTER_MODE=smart
export ROUTER_STRATEGY=balanced   # latency | cost | balanced
claudex

Telegram Gateway

Use Claudex through a Telegram bot. Each user gets an isolated session.

# configure (one time)
claudex telegram setup --token 123456:ABC --provider nvidia

# allow yourself
claudex telegram permit 987654321

# start the gateway
claudex telegram start

Or manage from inside the CLI:

/telegram setup --token 123456:ABC --provider nvidia
/telegram permit 987654321
/telegram status

Full guide: telegram-gateway/README.md


What Works

  • All tools: Bash, FileRead, FileWrite, FileEdit, Glob, Grep, WebFetch, WebSearch, Agent, MCP, LSP, NotebookEdit, Tasks
  • Real-time token streaming
  • Multi-step tool chains
  • Base64 and URL image inputs (vision models)
  • Slash commands: /commit, /review, /compact, /diff, /doctor, /provider, /telegram, etc.
  • Sub-agents via AgentTool
  • Persistent memory

What's Different from Upstream

  • No Anthropic extended thinking (OpenAI models use different reasoning)
  • No prompt caching (Anthropic-specific)
  • No Anthropic beta headers
  • Token output defaults to 32K — models that cap lower are handled gracefully
  • First-run skips Anthropic login if a provider profile is saved

Web Search

WebSearch is disabled by default for non-Anthropic providers. Set a Firecrawl key to enable it:

export FIRECRAWL_API_KEY=your-key

Free tier includes 500 credits. With this set, WebSearch works for all providers and WebFetch handles JS-rendered pages.


How the Shim Works

Claude Code Tool System
        │
        ▼
  Anthropic SDK interface (duck-typed)
        │
        ▼
  openaiShim.ts  ◄── translates formats
        │
        ▼
  OpenAI Chat Completions API
        │
        ▼
  Any compatible model

Model Quality Reference

| Model | Tool Calling | Code | Speed | |---|---|---|---| | GPT-4o | Excellent | Excellent | Fast | | Kimi K2 (NVIDIA) | Excellent | Excellent | Fast | | DeepSeek-V3 | Great | Great | Fast | | Gemini 2.0 Flash | Great | Good | Very Fast | | Llama 3.3 70B | Good | Good | Medium | | Mistral Large | Good | Good | Fast | | GPT-4o-mini | Good | Good | Very Fast | | Qwen 2.5 72B | Good | Good | Medium | | Models < 7B | Limited | Limited | Very Fast |


VS Code Extension

Install the bundled extension from vscode-extension/openclaude-vscode for one-click terminal launch and the Claudex Terminal Black theme.


Origin

Fork of the Claude Code source snapshot that became publicly accessible via an npm source map exposure on March 31, 2026. The original source is the property of Anthropic. This project is not affiliated with or endorsed by Anthropic.

License

Educational and research use. Original source subject to Anthropic's terms. Shim additions and new provider integrations are public domain.