npm package discovery and stats viewer.

Discover Tips

  • General search

    [free text search, go nuts!]

  • Package details

    pkg:[package-name]

  • User packages

    @[username]

Sponsor

Optimize Toolset

I’ve always been into building performant and accessible sites, but lately I’ve been taking it extremely seriously. So much so that I’ve been building a tool to help me optimize and monitor the sites that I build to make sure that I’m making an attempt to offer the best experience to those who visit them. If you’re into performant, accessible and SEO friendly sites, you might like it too! You can check it out at Optimize Toolset.

About

Hi, 👋, I’m Ryan Hefner  and I built this site for me, and you! The goal of this site was to provide an easy way for me to check the stats on my npm packages, both for prioritizing issues and updates, and to give me a little kick in the pants to keep up on stuff.

As I was building it, I realized that I was actually using the tool to build the tool, and figured I might as well put this out there and hopefully others will find it to be a fast and useful way to search and browse npm packages as I have.

If you’re interested in other things I’m working on, follow me on Twitter or check out the open source projects I’ve been publishing on GitHub.

I am also working on a Twitter bot for this site to tweet the most popular, newest, random packages from npm. Please follow that account now and it will start sending out packages soon–ish.

Open Software & Tools

This site wouldn’t be possible without the immense generosity and tireless efforts from the people who make contributions to the world and share their work via open source initiatives. Thank you 🙏

© 2026 – Pkg Stats / Ryan Hefner

ai-context-kit

v0.1.2

Published

Intelligent context management for AI coding agents. Load, lint, measure, and sync rules across Cursor, Claude Code, Copilot, and any AI tool.

Readme


You write a CLAUDE.md. Then someone adds .cursor/rules/. Then a teammate drops in an AGENTS.md. Then someone copies in a .cursorrules file from a blog post. Nobody removes the old ones.

Six months later your project has four context files that overlap, contradict each other, and dump 8,000 tokens of directory listings and "follow best practices" into every conversation. Your agent follows all of it. It gets slower. It gets confused. You blame the model.

An ETH Zurich study (February 2026) measured what actually happens when you give agents context files:

  • Auto-generated context files reduced task success compared to providing nothing
  • Human-written ones only improved accuracy by 4%
  • Inference costs jumped 20%+ from wasted tokens
  • Performance dropped on some models because agents got too obedient - following unnecessary instructions instead of solving the actual problem

I kept hitting this in my own projects, so I built ai-context-kit - a toolkit to treat context like a budget. Measure it, trim it, inject only what the current task needs.

import { loadRules, measure, lint, select } from "ai-context-kit";

const rules = await loadRules("./");

measure(rules, 4000); // what does your context cost?
lint(rules); // conflicts? duplicates? dead weight?
select(rules, {
  task: "fix auth bug", // only inject what matters
  budget: 2000, // stay within token budget
});

Quick Start

npm install ai-context-kit

Run the CLI on any project to see what you're actually injecting:

npx ai-context-kit measure
ai-context-kit measure - 6 rule file(s)

  Total: 4,821 tokens

  ############ 2,100 tokens (44%) - .cursor/rules/conventions.mdc
  ######## 1,200 tokens (25%) - CLAUDE.md
  ##### 890 tokens (18%) - .cursor/rules/api-patterns.mdc
  ## 340 tokens (7%) - AGENTS.md
  ## 180 tokens (4%) - .cursor/rules/testing.mdc
  # 111 tokens (2%) - .github/copilot-instructions.md

Then lint it:

npx ai-context-kit lint
ai-context-kit lint - 6 rule file(s)

  [!] .cursor/rules/conventions.mdc
      Rule is 2100 tokens. Consider splitting to keep each file under 2000 tokens.

  [x] CLAUDE.md
      Conflicts with AGENTS.md: "always use semicolons" vs "never use semicolons"

  [!] CLAUDE.md
      Duplicated line also found in .cursor/rules/conventions.mdc. Duplicates waste tokens.

  [i] AGENTS.md
      Contains vague instruction matching "follow best practices".
      Specific instructions produce better results than general advice.

  Score: 70/100 (FAILED)

That's the difference between guessing and knowing.


What's Different

| | Other approaches | ai-context-kit | | -------------- | ------------------------------------------------ | ------------------------------------------------------------ | | Context cost | Nobody measures it | Token count per file with budget check | | Conflicts | You find out when the agent does something weird | Detects contradictions across all files automatically | | Duplicates | Same rule in 3 files, 3x the tokens | Flagged and scored | | Task relevance | Every rule injected every time | select() picks only what matters for the current task | | Multi-tool | Locked to one IDE's format | Works across Cursor, Claude Code, Copilot, Windsurf, Cline | | CI | Hope for the best | lint exits with code 1 on errors. Drop it in your pipeline |


What This Answers

  1. How much context am I injecting? Token count per file, percentage breakdown, budget check
  2. Are my rules fighting each other? Conflict detection across all files and formats
  3. What's wasting tokens? Directory listings, duplicate content, vague advice
  4. Which rules matter for this task? Task-relevant selection with token budget

How It Works

ai-context-kit reads every context file format in the ecosystem, parses frontmatter, estimates token cost, and gives you tools to analyze and manage them.

| | | | ------------- | ----------------------------------------------------------------------------------------------------------------------------------- | | loadRules() | Auto-detects .cursor/rules/, .cursorrules, CLAUDE.md, AGENTS.md, copilot-instructions.md, .windsurfrules, .clinerules | | measure() | Token cost per rule, percentage of total, budget check | | lint() | Conflicts, duplicates, bloat, vague instructions, useless directory trees. Scores 0-100 | | select() | Picks rules relevant to the current task. Respects a token budget. alwaysApply rules first, then by relevance | | sync() | Single source of truth. Write once in .cursor/rules/, sync to CLAUDE.md, AGENTS.md, and the rest | | init() | Starter template with tips from the research |


API

loadRules(rootDir?)

const rules = await loadRules("./");
// Finds every context file in the project

const rules = await loadRules(".cursor/rules/");
// Or load from a specific directory

Returns RuleFile[] with parsed frontmatter, body, format, path, and token count.

measure(rules, budget?)

const report = measure(rules, 4000);

report.totalTokens; // 3847
report.overBudget; // false
report.rules; // sorted by size, each with tokens + percentage

lint(rules)

const report = lint(rules);

report.score; // 85/100
report.passed; // true (no errors, warnings don't fail)
report.issues; // array of { rule, path, severity, message }

What the linter catches:

| Rule | Severity | What it finds | | ------------------- | ------------- | ------------------------------------------------------------ | | token-budget | warning/error | Files over 2,000 tokens (warning) or 5,000 (error) | | empty-rule | warning | Files too short to do anything | | duplicate-content | warning | Same instruction repeated across files | | conflict | error | "always use X" in one file, "never use X" in another | | directory-listing | warning | 10+ line directory trees that agents don't need | | vague-instruction | info | "follow best practices", "write clean code", "be consistent" |

select(rules, options)

The core insight from the research: don't inject everything. Pick what matters.

const relevant = select(rules, {
  task: "fix auth bug in /api/auth",
  budget: 2000,
  tags: ["security", "api"],
  exclude: ["style"],
});

Scoring: alwaysApply: true in frontmatter gets highest priority. Then task words matched against file paths and content. Then tag matches. Budget is respected - highest-scored rules are included first until the budget runs out.

sync(options)

Write rules once, sync everywhere.

await sync({
  source: ".cursor/rules/",
  targets: ["CLAUDE.md", "AGENTS.md", ".github/copilot-instructions.md"],
});

Supports dryRun: true to preview changes without writing.

init(options?)

await init({ format: "cursor-rules" });
// Creates .cursor/rules/conventions.mdc with research-backed starter template

CLI

npx ai-context-kit lint                    # find issues
npx ai-context-kit lint --json             # machine-readable output
npx ai-context-kit measure                 # token cost breakdown
npx ai-context-kit measure --budget 4000   # check against budget
npx ai-context-kit sync --source .cursor/rules/ --target CLAUDE.md,AGENTS.md
npx ai-context-kit init                    # scaffold starter rules
npx ai-context-kit init --format claude-md

All commands support --path <dir> to point at a different project root. lint exits with code 1 on errors (warnings pass).


Use with Vercel AI SDK / LangChain / Custom Agents

This isn't just for Cursor. If you're building agents with Vercel AI SDK, LangChain, or your own framework, ai-context-kit solves the same problem: how much context are you stuffing into the system prompt, and is it helping or hurting?

import { loadRules, select } from "ai-context-kit";
import { generateText } from "ai";

const allRules = await loadRules("./rules");

const relevant = select(allRules, {
  task: userMessage,
  budget: 3000,
});

const systemPrompt = relevant.map((r) => r.body).join("\n\n");

const { text } = await generateText({
  model: openai("gpt-4o"),
  system: systemPrompt,
  prompt: userMessage,
});

Any framework that takes a system prompt string. Any rules stored as markdown files.


Supported Formats

| Format | File | Used by | | --------------- | --------------------------------- | -------------------- | | Cursor (modern) | .cursor/rules/*.mdc | Cursor IDE | | Cursor (legacy) | .cursorrules | Cursor IDE | | Claude Code | CLAUDE.md | Claude Code | | AGENTS.md | AGENTS.md | Cross-agent standard | | GitHub Copilot | .github/copilot-instructions.md | GitHub Copilot | | Windsurf | .windsurfrules | Windsurf | | Cline | .clinerules | Cline |

ai-context-kit detects the format from the file path. No configuration needed.


The ETH Zurich study tested both human-written and LLM-generated context files. Human-written ones were better, but only by 4%. The real problem isn't quality - it's volume. More context means more tokens consumed by instructions the agent doesn't need for the current task. The winning strategy is fewer, task-relevant rules, not better prose.

ai-context-kit uses a 4-character-per-token approximation. This is intentionally simple and fast. It's accurate enough for budgeting and comparison (GPT-4 averages ~4 chars/token for English text). If you need exact counts, pipe the output through tiktoken or your model's tokenizer.

Yes. npx ai-context-kit lint returns exit code 1 on errors, 0 on pass. Add it to your CI pipeline the same way you'd add eslint. The --json flag gives machine-readable output for custom reporting.


Tech Stack

| Component | Technology | | ------------ | ---------------------------------------------------------------------------------------------------------------------------------------------------- | | Language | TypeScript strict mode | | Testing | Vitest | | Bundler | tsup ESM + CJS | | Dependencies | Zero runtime dependencies |


Contributing

PRs welcome. Whether it's a new lint rule, a format detector, or a bug fix - check out the contributing guide.


Author

Made by ofershap

LinkedIn GitHub


README built with README Builder

License

MIT © Ofer Shapira