npm package discovery and stats viewer.

Discover Tips

  • General search

    [free text search, go nuts!]

  • Package details

    pkg:[package-name]

  • User packages

    @[username]

Sponsor

Optimize Toolset

I’ve always been into building performant and accessible sites, but lately I’ve been taking it extremely seriously. So much so that I’ve been building a tool to help me optimize and monitor the sites that I build to make sure that I’m making an attempt to offer the best experience to those who visit them. If you’re into performant, accessible and SEO friendly sites, you might like it too! You can check it out at Optimize Toolset.

About

Hi, 👋, I’m Ryan Hefner  and I built this site for me, and you! The goal of this site was to provide an easy way for me to check the stats on my npm packages, both for prioritizing issues and updates, and to give me a little kick in the pants to keep up on stuff.

As I was building it, I realized that I was actually using the tool to build the tool, and figured I might as well put this out there and hopefully others will find it to be a fast and useful way to search and browse npm packages as I have.

If you’re interested in other things I’m working on, follow me on Twitter or check out the open source projects I’ve been publishing on GitHub.

I am also working on a Twitter bot for this site to tweet the most popular, newest, random packages from npm. Please follow that account now and it will start sending out packages soon–ish.

Open Software & Tools

This site wouldn’t be possible without the immense generosity and tireless efforts from the people who make contributions to the world and share their work via open source initiatives. Thank you 🙏

© 2026 – Pkg Stats / Ryan Hefner

agentic-sdlc-wizard

v1.23.0

Published

SDLC enforcement for Claude Code — hooks, skills, and wizard setup in one command

Downloads

548

Readme

Claude Code SDLC Wizard

A self-evolving Software Development Life Cycle (SDLC) enforcement system for AI coding agents. Makes Claude plan before coding, test before shipping, and ask when uncertain. Measures itself getting better over time.

Install

Requires Claude Code (Anthropic's CLI for Claude).

Run from your terminal or from inside Claude Code (! prefix):

npx agentic-sdlc-wizard init

Then start (or restart) Claude Code — type /exit then claude to reload hooks. Setup auto-invokes on first prompt — Claude reads the wizard doc, scans your project, and generates bespoke CLAUDE.md, SDLC.md, TESTING.md, and ARCHITECTURE.md. No manual commands needed.

From GitHub (no npm needed):

npx github:BaseInfinity/agentic-ai-sdlc-wizard init

Manual: Download CLAUDE_CODE_SDLC_WIZARD.md to your project and tell Claude Run the SDLC wizard setup.

npx agentic-sdlc-wizard check        # Human-readable
npx agentic-sdlc-wizard check --json  # Machine-readable (CI-friendly)

Reports MATCH / CUSTOMIZED / MISSING / DRIFT for every installed file. Exits non-zero on MISSING or DRIFT — use in CI to catch setup regressions.

Check for content updates: Tell Claude Check if the SDLC wizard has updates — it reads CHANGELOG.md, shows what's new, and offers to apply changes.

Why Use This

You want Claude Code to follow engineering discipline automatically:

  • Plan before coding (not guess-and-check)
  • Write tests first (TDD enforced via hooks)
  • State confidence (LOW = ask user, don't guess)
  • Track work visibly (TaskCreate)
  • Self-review before presenting
  • Prove it's better (use native features unless you prove custom wins)

The wizard auto-detects your stack (package.json, test framework, deployment targets) and generates bespoke hooks + skills + docs. CI validates the generated assets; cross-stack setup-path E2E is on the roadmap.

What This Actually Is

Five layers working together:

Layer 5: SELF-IMPROVEMENT
  Weekly/monthly workflows detect changes, test them
  statistically, create PRs. Baselines evolve organically.

Layer 4: STATISTICAL VALIDATION
  E2E scoring with 95% CI (5 trials, t-distribution).
  SDP normalizes for model quality. CUSUM catches drift.

Layer 3: SCORING ENGINE
  7 criteria, 10/11 points. Claude evaluates Claude.
  Before/after wizard A/B comparison in CI.

Layer 2: ENFORCEMENT
  Hooks fire every interaction (~100 tokens).
  PreToolUse reminds Claude to write tests first.

Layer 1: PHILOSOPHY
  The wizard document. KISS. TDD. Confidence levels.
  Copy it, run setup, get a bespoke SDLC.

What Makes This Different

| Capability | What It Does | |---|---| | E2E scoring in CI | Every PR gets an automated SDLC compliance score (0-10) — measures whether Claude actually planned, tested, and reviewed | | Before/after A/B testing | Compares wizard changes against a baseline with 95% confidence intervals to prove improvements aren't noise | | SDP normalization | Separates "the model had a bad day" from "our SDLC broke" by cross-referencing external benchmarks | | CUSUM drift detection | Catches gradual quality decay over time — borrowed from manufacturing quality control | | Pre-tool TDD hooks | Before source edits, a hook reminds Claude to write tests first. CI scoring checks whether it actually followed TDD | | Self-evolving loop | Weekly/monthly external research + local CI shepherd loop — you approve, the system gets better |

How It Works

Think Iron Man: Jarvis is nothing without Tony Stark. Tony Stark is still Tony Stark. But together? They make Iron Man. This SDLC is your suit - you build it over time, improve it for your needs, and it makes you both better.

The dream: Mold an ever-evolving SDLC to your needs. Replace my components with native Claude Code features as they ship — and one day, delete this repo entirely because Claude Code has them all built in. That's the goal.

WIZARD FILE (CLAUDE_CODE_SDLC_WIZARD.md)
  - Setup guide, used once
  - Lives on GitHub, fetched when needed
        |
        | generates
        v
GENERATED FILES (in your repo)
  - .claude/hooks/*.sh
  - .claude/skills/*/SKILL.md
  - .claude/settings.json
  - CLAUDE.md, SDLC.md, TESTING.md, ARCHITECTURE.md
        |
        | validated by
        v
CI/CD PIPELINE
  - E2E: simulate SDLC task -> score 0-10
  - Before/after: main vs PR wizard
  - Statistical: 5x trials, 95% CI
  - Model-aware: SDP adjusts for external conditions

Self-Evolving System

| Cadence | Source | Action | |---------|--------|--------| | Weekly | Claude Code releases | PR with analysis + E2E test | | Weekly | Community (Reddit, HN) | Issue digest | | Monthly | Deep research, papers | Trend report |

Every update: regression tested -> AI reviewed -> human approved.

E2E Scoring

Like evaluating scientific method adherence - we measure process compliance:

| Criterion | Points | Type | |-----------|--------|------| | TodoWrite/TaskCreate | 1 | Deterministic | | Confidence stated | 1 | Deterministic | | Plan mode | 2 | AI-judge | | TDD RED | 2 | Deterministic | | TDD GREEN | 2 | AI-judge | | Self-review | 1 | AI-judge | | Clean code | 1 | AI-judge |

40% deterministic + 60% AI-judged. 5 trials handle variance.

Model-Adjusted Scoring (SDP)

| Metric | Meaning | |--------|---------| | Raw | Actual score (Layer 2: SDLC compliance) | | SDP | Adjusted for model conditions | | Robustness | How well SDLC holds up vs model changes |

  • Robustness < 1.0 = SDLC is resilient (good!)
  • Robustness > 1.0 = SDLC is sensitive (investigate)

Tests Are The Building Blocks

Tests aren't just validation - they're the foundation everything else builds on.

  • Tests >= App Code - Critique tests as hard (or harder) than implementation
  • Tests prove correctness - Without them, you're just hoping
  • Tests enable fearless change - Refactor confidently

Official Plugin Integration

| Plugin | Purpose | Scope | |--------|---------|-------| | claude-md-management | Required - CLAUDE.md maintenance | CLAUDE.md only | | claude-code-setup | Recommends automations | Recommendations | | code-review | Local self-review and PR review (optional) | Local + PRs |

Prove It's Better

Don't reinvent the wheel. Use native/built-in features UNLESS you prove your custom version is better. If you can't prove it, delete yours.

  1. Test the native solution — measure quality, speed, reliability
  2. Test your custom solution — same scenario, same metrics
  3. Compare side-by-side
  4. Native >= custom? Use native. Delete yours.
  5. Custom > native? Keep yours. Document WHY. Re-evaluate when native improves.

This applies to everything: native commands vs custom skills, framework utilities vs hand-rolled code, library functions vs custom implementations.

How This Compares

This isn't the only Claude Code SDLC tool. Here's an honest comparison:

| Aspect | SDLC Wizard | everything-claude-code | claude-sdlc | |--------|------------|----------------------|-------------| | Focus | SDLC enforcement + measurement | Agent performance optimization | Plugin marketplace | | Hooks | 3 (SDLC, TDD, instructions) | 12+ (dev blocker, prettier, etc.) | Webhook watcher | | Skills | 4 (/sdlc, /setup, /update, /feedback) | 80+ domain-specific | 13 slash commands | | Evaluation | 95% CI, CUSUM, SDP, Tier 1/2 | Configuration testing | skilltest framework | | CI Shepherd | Local CI fix loop | No | No | | Auto-updates | Weekly CC + community scan | No | No | | Install | npx agentic-sdlc-wizard init | npm install | npm install | | Philosophy | Lightweight, prove-it-or-delete | Scale and optimization | Documentation-first |

Our unique strengths: Statistical rigor (CUSUM + 95% CI), SDP scoring (model quality vs SDLC compliance), CI shepherd loop, Prove-It A/B pipeline, comprehensive automated test suite, dogfooding enforcement.

Where others are stronger: everything-claude-code has broader language/framework coverage. claude-sdlc has webhook-driven automation. Both have npm distribution.

The spirit: Open source — we learn from each other. See COMPETITIVE_AUDIT.md for details.

Documentation

| Document | What It Covers | |----------|---------------| | ARCHITECTURE.md | System design, 5-layer diagram, data flows, file structure | | CI_CD.md | All 4 workflows, E2E scoring, tier system, SDP, integrity checks | | SDLC.md | Version tracking, enforcement rules, SDLC configuration | | TESTING.md | Testing philosophy, test diamond, TDD approach | | CHANGELOG.md | Version history, what changed and when | | CONTRIBUTING.md | How to contribute, evaluation methodology |

Contributing

PRs welcome. See CONTRIBUTING.md for evaluation methodology and testing.