npm package discovery and stats viewer.

Discover Tips

  • General search

    [free text search, go nuts!]

  • Package details

    pkg:[package-name]

  • User packages

    @[username]

Sponsor

Optimize Toolset

I’ve always been into building performant and accessible sites, but lately I’ve been taking it extremely seriously. So much so that I’ve been building a tool to help me optimize and monitor the sites that I build to make sure that I’m making an attempt to offer the best experience to those who visit them. If you’re into performant, accessible and SEO friendly sites, you might like it too! You can check it out at Optimize Toolset.

About

Hi, 👋, I’m Ryan Hefner  and I built this site for me, and you! The goal of this site was to provide an easy way for me to check the stats on my npm packages, both for prioritizing issues and updates, and to give me a little kick in the pants to keep up on stuff.

As I was building it, I realized that I was actually using the tool to build the tool, and figured I might as well put this out there and hopefully others will find it to be a fast and useful way to search and browse npm packages as I have.

If you’re interested in other things I’m working on, follow me on Twitter or check out the open source projects I’ve been publishing on GitHub.

I am also working on a Twitter bot for this site to tweet the most popular, newest, random packages from npm. Please follow that account now and it will start sending out packages soon–ish.

Open Software & Tools

This site wouldn’t be possible without the immense generosity and tireless efforts from the people who make contributions to the world and share their work via open source initiatives. Thank you 🙏

© 2026 – Pkg Stats / Ryan Hefner

zodiac-skills

v2.3.0

Published

Behavioral compiler for LLMs — 12 skills that break specific model defaults. Cross-model, verified against real-world benchmarks.

Readme

zodiac-skills

A behavioral compiler for LLMs. Twelve SKILL.md files — each one rewires how a model thinks about a specific class of problem. No fine-tuning, no RAG, no vector database. Just constraints that make LLMs do what they won't do by default: disagree with themselves, admit uncertainty, check before claiming absence, and produce findings that survive contact with reality.

Works with Claude Code, Codex, Cursor, Gemini CLI, and any tool that reads SKILL.md.

npx zodiac-skills

Proof

We ran the zodiac on real targets and verified every finding against published expert analysis. Not cherry-picked examples — full verification, including what we got wrong.

EU AI Act — Full Regulation (573KB, 113 articles)

8 animals. 2 Rabbit-orchestrated passes. 40 findings.

| Metric | Result | |--------|--------| | Findings contradicted by published sources | 0 / 40 | | Findings confirmed by legal scholarship | 17 / 40 | | Findings ahead of published discourse | 2 (in-context learning gap, RAG blind spot) | | Rat's prediction of notified body shortage | Confirmed 6 months later by industry bodies | | Dog's product-safety → fundamental-rights drift finding | Independently published in Common Market Law Review | | Rat's emotion recognition scope-creep finding | Confirmed AND exceeded — Commission guidelines widened the exception further than the analysis predicted |

Full showcase — includes verification methodology and per-finding source mapping.

Zodiac vs. Vanilla — Controlled A/B Test

Same model (Claude Opus 4.6). Same input (Trump AI Legislative Framework, 2 days old — neither run could rely on memorized analysis). Same request: rigorous critical analysis.

| | Vanilla | Zodiac | |---|---|---| | Unique findings | 5 (that zodiac missed) | 10 (that vanilla missed) | | Falsifiable claims | Low — qualitative judgments | Higher — Delete Probe, Precision Probe, Inversion Attack | | Confidence calibration | None — uniform implicit confidence | Explicit scores per finding | | Hedging | Moderate — "has some merit," "arguably" | Binary verdicts — Burned or Survived, no middle ground | | Positive findings | Afterthought | Mandatory — forced genuine search for what works | | Cross-referencing | None — sections are independent | Cross-animal tensions surfaced |

The vanilla analysis was a good essay. The zodiac was a diagnostic instrument.

Full comparison — includes finding-by-finding breakdown and honest accounting of where vanilla outperformed zodiac.

More Showcases

| Target | Animals | Key result | |--------|---------|------------| | EU AI Act Article 5 | Rooster, Monkey, Tiger, Pig, Dragon | Every finding verified against Commission guidelines, FPF, Verfassungsblog, law firm analyses | | Trump AI Framework | Rooster, Monkey, Tiger, Pig, Dragon | Delete Probe: "remove 4 of 6 prongs — nothing changes." Confirmed by CSIS, Sullivan & Cromwell | | Tesla Q4 2025 Earnings | Rooster, Rat, Pig | Earnings call analysis — claims verified against SEC filings | | Apple Q1 2026 Earnings | Rooster, Rat, Pig | Revenue claims fact-checked, Rat traced margin consequence chains | | Artisan Pitch Deck | Rooster, Monkey, Pig | Seed deck stress-tested — unit economics and moat claims audited | | darktable Architecture | Rooster, Monkey, Tiger | Open-source codebase architecture review — anti-fabrication rule caught 3 false absence claims | | Spine Swarm YC Launch | Monkey, Tiger, Rooster, Snake | YC S23 launch analyzed — technical claims vs. demo reality | | Supabase Architecture | Monkey, Tiger, Rooster, Snake | 13-service stack stress-tested — RLS default-off confirmed by CVE + security research | | Anthropic Agent Skills Blog | Rooster, Rat, Pig | Blog post claims audited against actual SDK capabilities |

The Animals

Each animal is a single SKILL.md file that breaks one specific LLM default. No dependencies. No supporting files. One file, one behavioral change.

| Animal | Breaks | What it does | |--------|--------|-------------| | 猴 Monkey | Agreeableness | Stress-tests assumptions from 9 angles. Survived: yes/no | | 牛 Ox | Pattern matching | Questions whether the chosen pattern is warranted. Warranted: yes/no | | 虎 Tiger | Premature convergence | Attacks the solution from multiple angles to find fatal flaws. Burned: yes/no | | 蛇 Snake | Scope creep | Cuts everything that hasn't earned its place. Earned: yes/no | | 鼠 Rat | Linear thinking | Maps second and third-order consequences. Contained: yes/no | | 狗 Dog | Context drift | Detects when a project has wandered from original intent. Aligned: yes/no | | 猪 Pig | Political hedging | Says what everyone's thinking but nobody will say. Hedged: yes/no | | 兔 Rabbit | Audience-blind output | Orchestrates other animals, shapes output for the reader. | | 龙 Dragon | Short-term thinking | Analyzes decisions for reversibility, lock-in, and compounding. Farsighted: yes/no | | 马 Horse | Analysis paralysis | Cuts through overthinking to identify what's actually blocking. Clear: yes/no | | 羊 Goat | Convergent thinking | Explores alternatives nobody considered. Fertile: yes/no | | 鸡 Rooster | Epistemic recklessness | Fact-checks claims and audits evidence. Verified: yes/no |

The binary success mechanic (Survived: yes is a valid exit) prevents confirmation bias — the animals don't manufacture findings when there's nothing to find.

How It Works

Decision Policy is load-bearing, persona is flavor. Each SKILL.md has operational rules — what to distrust, what evidence to require, when the anti-scope kicks in ("you don't fix things," "you don't propose alternatives"). These rules drive behavior across any model. The character voice (enthusiastic Monkey, methodical Ox, silent Snake) adds tone but doesn't carry the instruction.

Named technique arsenals with "never repeat." Without named techniques, LLMs gravitate to 2-3 analytical patterns. Each animal has a named arsenal (Assumption Flip, Hostile Input, Delete Probe, Scale Shift...) that forces variety. The Monkey uses 9 distinct techniques; every other animal uses 5.

Anti-fabrication on absence claims. Before an animal claims something doesn't exist, it must state where it looked. "I did not find X in Y" is honest. "There is no X" is a claim requiring verification. This rule was added after the Monkey's three highest-confidence findings on darktable were factually wrong — they asserted absence without checking primary docs.

The Rabbit orchestrates. It picks which animals to run, calibrates finding counts to target density (fewer high-quality findings on thin targets), decomposes document-sized targets into sections, and synthesizes cross-animal output into a shaped deliverable. Raw outputs go to file; conversation stays lean.

Install

npx zodiac-skills

Interactive prompts for tool and scope:

  Which tool?
    1) Claude Code
    2) Codex
    3) Both

  Where should skills be installed?
    1) User      (available everywhere)
    2) Project   (current directory only)

Or skip prompts:

npx zodiac-skills --claude --user              # Claude Code, user-level
npx zodiac-skills --codex --project            # Codex, project-level
npx zodiac-skills --claude --codex --user      # Both tools, user-level
npx zodiac-skills install monkey tiger snake   # Specific animals
npx zodiac-skills list                         # Check what's installed
npx zodiac-skills uninstall --all              # Remove all

Usage

Invoke by name:

/monkey — chaos check this architecture
/tiger — attack this migration plan
/snake — cut this PRD to MVP
/rat — map consequences of this API change
/rooster — verify these performance claims
/rabbit — review this proposal (picks animals, shapes output)

The Rabbit is the front door for multi-animal analysis. Give it a target, it picks the right animals, runs them, and delivers a synthesized result shaped for your audience:

/rabbit — review this 50-page architecture doc

Rabbit decomposes the target into sections, selects animals per-section, calibrates finding counts to density, and outputs ~1200 words to conversation with full raw outputs saved to rabbit-output/.

License

MIT