vexp-cli
v2.0.12
Published
Vexp — Context Engine for AI Coding Agents. Pre-indexes your codebase into a dependency graph and delivers ranked context to any MCP-compatible agent. 58% lower cost per task, 90% fewer tool calls (SWE-bench Verified). Works with Claude Code, Cursor, Copi
Maintainers
Keywords
Readme
vexp — Context Engine for AI Coding Agents
58% Less Cost. 90% Fewer Tool Calls. 100% On Your Machine.
vexp is a local-first context engine that gives AI coding agents the right code context — automatically. One install, and every agent gets smarter answers with fewer tokens.
Works with 14+ AI coding agents and 30 programming languages out of the box.
→ SWE-bench Verified: 73% pass@1 at $0.67/task
Why vexp?
AI coding agents waste more than half their budget exploring codebases before writing a single line of code — 23 tool calls on average just to orient themselves. vexp pre-indexes your project into a dependency graph and delivers only the code that matters, ranked by relevance.
- 23 tool calls → 2 — the agent stops guessing and starts coding
- $0.78 → $0.33 per task — measured on SWE-bench Verified
- 117s → 74s per task — less context to process = faster answers
- Local-first — your code never leaves your machine
- Zero config — auto-detects your agents and configures MCP
- Universal — works with any agent that supports MCP or custom rules
- Optional local AI — unlock +30% extra savings with
vexp-devmind, a ~3.5 GB model that runs 100% on your machine. GPU-accelerated on NVIDIA (CUDA) and Apple Silicon (Metal). Opt in with one command.
Supported AI Coding Agents
| Agent | Auto-Configured | |-------|:-:| | Claude Code | ✓ | | Cursor | ✓ | | Windsurf | ✓ | | GitHub Copilot | ✓ | | Continue.dev | ✓ | | Augment | ✓ | | Zed | ✓ | | Codex (OpenAI) | ✓ | | Cline | ✓ | | Aider | ✓ | | Kilo Code | ✓ | | Kiro | ✓ | | Antigravity | ✓ | | Opencode | ✓ |
Your agents don't compete with each other — they all share the same context layer.
Supported Languages (30)
TypeScript, JavaScript, Python, Go, Rust, Java, C#, C, C++, Ruby, Kotlin, Scala, Swift, Dart, Elixir, Haskell, OCaml, Lua, R, PHP, Zig, HCL/Terraform, Objective-C, Bash/Shell, Dockerfile, Clojure, F#, SQL, HTML/CSS, YAML/JSON
Quick Start
npm install -g vexp-cli
vexpThat's it. The interactive CLI guides you through setup, indexing, and agent configuration.
What You Get
- Semantic + graph-ranked context — finds the right symbols even when keywords don't match
- Intent detection in 15 languages — English, Italian, French, German, Spanish, Portuguese, Russian, Chinese, Japanese, Korean, Arabic, Hindi, Turkish, Dutch, Polish
- Cross-session memory — your agent picks up exactly where it left off, even days later
- Stale-knowledge alerts — when code changes, outdated memory is automatically flagged
- Pattern & anti-pattern detection — recurring changes surface as guidelines; dead-end explorations get flagged early
- Change awareness — files frequently modified together are automatically linked
- Security filter —
.env, credentials, and SSH keys are excluded automatically
Platforms
Native binaries — no runtime dependencies:
| Platform | Architecture | GPU Acceleration | |----------|:-:|:-:| | Linux | x64, ARM64 | CUDA (NVIDIA) | | macOS | x64, ARM64 (Apple Silicon) | Metal (built-in) | | Windows | x64 | CUDA (NVIDIA) |
GPU acceleration applies to the optional local AI layer. Vexp's core context engine runs on any modern CPU.
VS Code Extension
For the best experience, install the vexp VS Code extension — it manages the daemon automatically and adds a sidebar dashboard with token savings tracking.
Local AI — Optional +30% Savings
Turn on vexp-devmind — a code-specialised AI model that runs entirely on your machine — for an extra +30% token savings on top of the baseline.
- 100% local — the model never calls out. Your code stays on your hardware.
- GPU-accelerated — Metal on Apple Silicon, CUDA on NVIDIA. Auto-detected.
- One-command install — ~3.5 GB one-time download. Minimum 6 GB RAM.
- Always up to date —
--check-updatescompares your local version to the published release without downloading anything.
Sharper context ranking, higher-quality session memory, smarter prompt preprocessing. All without sending a single byte to any server.
vexp setup-llm --install # ~3.5 GB one-time download
vexp setup-llm --check-updates # no-op if already up to date
vexp setup-llm --status # show current config + hardware
vexp setup-llm --disable # turn it off any timeLinks
License
MIT
