commands-com
v0.2.1
Published
Local-first OSS CLI for review cycles and code quality workflows
Maintainers
Readme
Commands.com CLI
One command to improve your codebase until it earns an A.
npx commands-com quality --until ACommands.com fans out across local AI CLIs, synthesizes a code-quality grade, plans non-overlapping fixes, runs implementers in parallel, re-reviews, and keeps going until the target grade is reached or the safety cap stops it.
Set It. Walk Away.
Most agent tools want you babysitting every prompt. Commands.com runs a closed loop until your code earns the grade you asked for — or hits the safety cap.
audit -> synthesize -> plan fixes -> run implementers -> validate -> re-auditEverything lands in plain files under .commands-com/runs/: prompts, provider
outputs, preflight checks, task status, resume state, Markdown reports, and the
final summary. No daemon. No hosted state. Nothing you can't cat.
The headline move:
npx commands-com quality --until A --test "npm test"Walk away. Come back to a scored final report.
Quick Start
Try it without installing:
npx commands-com quality --until AUse a deterministic local mock provider:
npx commands-com quality --provider mock --until ASave provider defaults:
commands-com init --providers allRun a normal review:
commands-com review "review the current diff" --changed
commands-com review "fix the failing tests" --fix --test "npm test"Resume after a provider outage, capacity error, or interrupted shell:
commands-com quality --resume latest
commands-com review --resume <run-id>Inspect the trail:
commands-com runs list
commands-com runs show <run-id>Requires Node.js 20.19 or newer.
What It Does
- Quality to A:
quality --until Akeeps cycling until the synthesized A-F score reaches the target. - Scored review:
review --until Auses the same stop shape for code review loops. - Parallel fan-out: provider × area and provider × reviewer jobs run together by default.
- One synthesized answer: provider findings are deduped into a final score, synopsis, and prioritized issues.
- Multi-implementer fixes: the orchestrator chooses tasks up to
--max-implementers, then workers run non-overlapping fixes. - Provider fallback: transient capacity and stream failures retry, then fall through to the next provider.
- Resume-safe runs:
run-state.jsonlets--resumecontinue from the previous cycle. - Plain local files: no daemon, no hosted project state, no mystery database.
Providers
review and quality default to --providers all, which uses every available
real provider CLI in order: Codex, Claude, Gemini. The built-in mock provider
keeps tests and demos fast.
commands-com doctor --ping
commands-com quality --providers codex,claude,gemini --until A
commands-com quality --provider claude --area maintainabilityEnvironment defaults work too:
COMMANDS_COM_PROVIDERS=codex,claude commands-com quality --until A
COMMANDS_COM_PROVIDER=codex COMMANDS_COM_MODEL=gpt-5.5 commands-com review "review this"Safety
Report-only is the default. File edits require --fix or --until.
Commands.com refuses to edit a dirty current worktree unless you pass
--allow-dirty or isolate edits in a git worktree:
commands-com quality --until A --worktree --test "npm test"The --test command runs through your shell with your environment, so shell metacharacters in its value are interpreted by the shell.
Fix loops also write preflight.json, run-state.json, final-summary.json,
and final-report.md so failed or interrupted runs are understandable.
Commands
| Command | Use it for |
| --- | --- |
| quality | Score and improve code quality. |
| review <objective> | Review a diff or objective, optionally with fixes. |
| room <type> <objective> | Run a focused multi-perspective room. |
| rooms list | See available rooms. |
| runs list / runs show <id> | Inspect local run artifacts. |
| doctor | Check provider CLI availability. |
| init | Save local provider defaults. |
Built-in rooms include security, architecture, performance, bug-hunt,
test-plan, implementation-plan, product-spec, docs,
release-readiness, and codebase-research.
commands-com room security "audit the local runtime"
commands-com room implementation-plan "ship the v1 unlock"Flags
Run commands-com help for full descriptions. The short version:
Common Flags
--cwd, --json
Review Flags
--providers, --provider, --model, --changed, --until, --fix, --resume,
--worktree, --base-ref, --allow-dirty, --keep-worktree,
--max-cycles, --max-implementers, --stall-cycles, --reviewers,
--parallel, --serial, --retries, --test, --timeout-ms,
--fail-on-issues
Quality Flags
--providers, --provider, --model, --changed, --until, --fix,
--resume, --worktree, --base-ref, --allow-dirty, --keep-worktree,
--max-cycles, --max-implementers, --stall-cycles, --area,
--parallel, --serial, --retries, --test, --timeout-ms,
--fail-on-issues
Room Flags
--provider, --model, --changed, --participants, --parallel,
--no-synthesis, --retries, --timeout-ms
Doctor Flags
--providers, --provider, --model, --ping, --timeout-ms
Init Flags
--providers, --provider, --model
Runs Flags
--limit
Run Layout
Every run writes a folder like:
.commands-com/runs/<run-id>/
metadata.json
preflight.json
run-state.json
context.md
prompts/
cycle-1/
synthesis.md
areas/ or reviewers/
implementation-plan.md
tasks/
implementation.md
test.log
code-quality.md or review-cycle.md
final-summary.json
final-report.mdDevelopment
npm install
npm run validateThe mock provider is part of the product. It keeps tests and docs usable on
machines without AI provider CLIs installed.
See SECURITY.md for the --fix and --test trust model, and
CONTRIBUTING.md for project guardrails.
License
MIT, see LICENSE.
