npm package discovery and stats viewer.

Discover Tips

  • General search

    [free text search, go nuts!]

  • Package details

    pkg:[package-name]

  • User packages

    @[username]

Sponsor

Optimize Toolset

I’ve always been into building performant and accessible sites, but lately I’ve been taking it extremely seriously. So much so that I’ve been building a tool to help me optimize and monitor the sites that I build to make sure that I’m making an attempt to offer the best experience to those who visit them. If you’re into performant, accessible and SEO friendly sites, you might like it too! You can check it out at Optimize Toolset.

About

Hi, 👋, I’m Ryan Hefner  and I built this site for me, and you! The goal of this site was to provide an easy way for me to check the stats on my npm packages, both for prioritizing issues and updates, and to give me a little kick in the pants to keep up on stuff.

As I was building it, I realized that I was actually using the tool to build the tool, and figured I might as well put this out there and hopefully others will find it to be a fast and useful way to search and browse npm packages as I have.

If you’re interested in other things I’m working on, follow me on Twitter or check out the open source projects I’ve been publishing on GitHub.

I am also working on a Twitter bot for this site to tweet the most popular, newest, random packages from npm. Please follow that account now and it will start sending out packages soon–ish.

Open Software & Tools

This site wouldn’t be possible without the immense generosity and tireless efforts from the people who make contributions to the world and share their work via open source initiatives. Thank you 🙏

© 2026 – Pkg Stats / Ryan Hefner

agilai

v1.5.0

Published

Universal AI agent framework with MCP integration - Chat naturally, get professional results

Readme

Agilai

License: MIT Node.js Version npm version Package Size

🚀 From Idea to Implementation in Under Five Minutes

Professional specs • Visual concepts • Architecture docs • Ready-to-build stories

All through natural conversation. No methodology training. No prompt engineering.

Quick StartSee It In ActionDocumentation


Stop Fighting Your Tools. Start Building.

The Problem: You have a brilliant idea. But before you write a single line of code, you need:

  • A product requirements document (so the team knows what to build)
  • Technical architecture (so it's built right)
  • Visual design concepts (so it looks professional)
  • User stories (so developers know what to implement)

Traditional approach? Days of writing specs, creating mockups, and coordinating team meetings.

AI tools today? Hours of prompt engineering just to get structured output.

There has to be a better way.

The Agilai Way: Talk Naturally. Ship Professionally.

Imagine describing your project idea in plain English and getting back:

Complete PRD with user stories, acceptance criteria, and technical requirements 🎨 Visual concept explorations for every screen in your journey 🏗️ Technical architecture with file structure, tech stack, and implementation guidance 📋 Development-ready stories that developers can pick up immediately

No jargon. No templates. No methodology training.

Just conversation → professional deliverables → ready to build.

How It Works

1. Quick Lane (Under 5 Minutes)

  • Natural conversation about your idea
  • Auto-generated specs, architecture, and visual prompts
  • Implementation-ready output in docs/

2. Complex Lane (Enterprise-Grade)

  • Deep-dive conversational discovery
  • Visual journey mapping with CSS token extraction
  • Multi-phase QA-ready documentation

3. Conversational UI Designer (NEW 🆕)

  • Guide users through their UI journey step-by-step
  • Extract design tokens from reference URLs via Chrome DevTools
  • Generate per-screen visual prompts for Google Nano Banana
  • Capture decisions with journey context and CSS specifications

What Makes Agilai Different

🎯 Zero Prompt Engineering

Other tools: "Act as a senior product manager with 10 years experience in SaaS..." Agilai: "I want to build a task management app"

🎨 Visual-First Development

Traditional: Write specs → Hire designer → Wait for mockups → Start coding Agilai: Describe idea → Get visual concepts in minutes → Architecture adapts automatically

🏃 Two Speeds, One System

Quick Lane: Ship specs in 5 minutes Complex Lane: Enterprise depth when needed Seamlessly switch between both

🔒 Local-First, Private

No API calls to external services Runs through Claude CLI, Codex CLI, or OpenCode Your data never leaves your machine

Key Features

  • Conversational Discovery - Multi-turn UI journey mapping that extracts design intent naturally
  • Chrome MCP Integration - Automatically pull CSS tokens, color palettes, and typography from reference URLs
  • Per-Screen Visual Prompts - Tailored Nano Banana prompts for each step in your user journey
  • Dual-Lane Intelligence - Quick Lane for speed, Complex Lane for depth
  • Professional Output - Industry-standard PRDs, architecture docs, and user stories
  • Built on BMAD-METHOD™ - Battle-tested agile processes, invisible to users
  • Local & Private - Works with your existing CLI tooling, no API costs

NEW: Conversational UI Designer - The only AI workflow that guides you through visual discovery the way a real designer would: journey mapping → inspiration intake → CSS extraction → per-screen prompts. See it in action →

Quick Start

Prerequisites

  • Node.js ≥ 20.10.0
  • npm ≥ 9.0.0
  • One chat CLI installed: Claude CLI, Codex CLI, or OpenCode CLI

Installation

# One command does everything
npx agilai@latest start

# Optional: Specify assistant and provider
npx agilai@latest start --assistant=claude --glm

That's it! This command:

  1. Creates project structure
  2. Installs dependencies
  3. Builds MCP server
  4. Launches chosen chat interface

Complete Installation Guide - All installation options

Natural Conversation Example

You: "I want to build an app for tracking family chores"

AI: "Great! Let's figure this out together.
     Who will be using this app?"

You: "Parents and kids, ages 8 and up"

AI: "Perfect. What does success look like?"

[... natural conversation continues ...]

AI: "📋 Here's what I understand:
     Project: Family Chore Tracking App
     Users: Parents & children (8+)
     Timeline: 3 months
     Platform: Mobile-first

     Is this correct? (y/n)"

Behind the scenes, Agilai generates:

  • Project brief (analyst phase)
  • Product requirements document (PM phase)
  • Technical architecture (architect phase)
  • User stories (scrum master phase)
  • Implementation guidance (dev phase)

All organized in docs/ folder, ready to use.

Complete Examples - Detailed walkthrough examples

Conversational UI Designer Workflow

The first AI workflow that designs UIs the way a real designer would: through conversation.

🎨 How Real Designers Work

  1. Discover the journey - "Walk me through how users will flow through your app"
  2. Deep-dive each step - "What's the user's mindset on this screen? What emotion should they feel?"
  3. Gather inspiration - "Share a reference URL - I'll extract the colors, fonts, and spacing"
  4. Define visual language - "Let's lock in your palette, typography, and layout system"
  5. Generate concepts - "Here are tailored prompts for each screen with all the context"

Agilai's UI Designer Liaison does exactly this.

Quick Lane (Auto Mode)

You: "I want to build a task management app for remote teams"

AI: "✅ Generated PRD, architecture, and stories
     📋 Quick Lane also created:

     docs/ui/ui-designer-screen-prompts.md

     I've inferred your user journey from the PRD:
     1. Browse tasks
     2. Search & filter
     3. Create task
     4. Task details
     5. Assign & collaborate

     Each screen has a tailored visual concept prompt ready for
     Google Nano Banana (Gemini 2.5 Flash Image).

     Copy the prompts to https://aistudio.google.com and get
     3 concept variations per screen!"

Time: Under 5 minutes, fully automatic

Complex Lane (Conversational Mode)

You: "@ui-designer-liaison"

Nana: "Hi! I'm Nana, your UI designer liaison. Ready to craft your
       visual journey?

       🎨 Existing inspiration (URLs, reference designs)
       ✨ From scratch (I'll guide you)
       🔀 Both

       Which approach works best?"

You: "I have Linear.app as inspiration"

Nana: "Perfect! I notice you want to use Linear.app for reference.

       To automatically extract colors, typography, and CSS tokens,
       I recommend enabling Chrome DevTools MCP. It's optional but
       saves manual work.

       Would you like to activate it?"

You: "Yes, it's enabled"

Nana: "Excellent! Let me extract the visual elements...

       🎨 Color Palette:
       - Primary: #5E6AD2 (purple)
       - Accent: #3D9970 (green)
       - Neutral: #6B7280 (gray)

       ✍️ Typography:
       - Headings: 'Inter', sans-serif (600 weight)
       - Body: 'Inter', sans-serif (400 weight)

       📏 Spacing: 4px, 8px, 12px, 16px, 24px, 32px

       What elements should we KEEP vs AVOID from Linear?"

You: "Keep the minimal sidebar and generous spacing. Avoid the complex
      command palette."

Nana: "Got it! Now walk me through your ideal user journey from first
       touch to success..."

[... 6-stage conversational flow continues ...]

Nana: "✅ Visual concept prompts generated!

       📄 docs/ui/ui-designer-screen-prompts.md

       5 screens mapped:
       1. Browse tasks
       2. Search & filter
       3. Create task
       4. Task details
       5. Collaborate

       Each prompt includes:
       - Journey context (position, adjacent steps)
       - Persona mindset & goals
       - CSS tokens from Linear.app
       - Your visual preferences

       Copy to Google AI Studio and generate concepts!"

Time: 10-15 minutes, rich context, CSS extraction

What You Get

Per-Screen Prompts like this:

You are Google Nano Banana (Gemini 2.5 Flash Image).
Render 3 concept options for the **Search & Filter** step of TaskFlow Pro.

## Context
- Journey position: Step 2 of 5 (Early exploration phase)
- Previous step: Browse tasks
- Next step: Task details
- Persona mindset: User with specific criteria looking to narrow task list
- Screen goal: Filter tasks by status, assignee, due date efficiently
- Emotional tone: Focused, in-control

## Visual System
### Brand Palette
#5E6AD2, #3D9970, #6B7280

**CSS Tokens:**
--color-primary: #5E6AD2;
--color-accent: #3D9970;
--font-heading: 'Inter', sans-serif;
--space-base: 4px;
--space-md: 12px;

### Typography
- Heading: Inter 600
- Body: Inter 400
- Scale: 14px body, 18px subtitle, 24px heading

[... complete prompt with all context ...]

Chrome MCP Integration

Extract design tokens from any URL:

  1. Share a reference URL (Linear, Notion, Stripe, etc.)
  2. Chrome DevTools MCP automatically captures:
    • Color palette (hex codes)
    • Typography (font families, weights, sizes)
    • Spacing tokens (margins, padding)
    • CSS custom properties
  3. Tokens embedded in every screen prompt
  4. Developers get CSS-ready specifications

Log Your Selection

You: "*log-selection"

Nana: "Let's record your selected visual concept!

       Which concept did you choose?"

You: "Concept 2 - the clean minimal with purple accents"

Nana: "Perfect! I'll log:
       ✅ Visual characteristics
       ✅ Journey context (5 screens)
       ✅ CSS tokens from Linear.app
       ✅ Per-screen prompts used
       ✅ Asset locations

       Saved to: docs/ui/ui-designer-explorations.md

       This informs:
       - Frontend architecture
       - Component library setup
       - Design system documentation"

Benefits

🚀 10x Faster Than Traditional Design

Traditional: Designer creates mockups (days) → Revisions (days) → Developer handoff Agilai: Conversation (10 min) → Concepts (minutes) → CSS tokens included

🎯 Context-Aware Prompts

Generic tools: Single prompt for entire app Agilai: Tailored prompt per screen with journey context

🔧 Developer-Ready Output

Other tools: Images only Agilai: Images + CSS tokens + design system documentation

💡 Inspiration Made Actionable

Copy visual elements: Manual color picking from screenshots Agilai: Automatic CSS extraction with Chrome DevTools MCP

Try It Now

Quick Lane (Auto):

npx agilai@latest start
# Describe your UI project
# Get visual prompts automatically

Complex Lane (Conversational):

npx agilai@latest start
# Activate: @ui-designer-liaison
# Run: *discover-journey
# Full 6-stage designer conversation

Complete UI Designer Guide - Full workflow documentation → Using Google Nano Banana - Google AI Studio integration

Extend with Any Tool

Add powerful integrations through natural conversation:

You: "I need GitHub integration"

AI: "I can help with that! I'll set up GitHub access.
     What's your GitHub personal access token?"

You: "ghp_****"

AI: "✓ GitHub integration configured!
     Please restart your chat session for changes to take effect."

[You restart the CLI]

You: "Create an issue for the login bug"

AI: "✓ Created issue #42: 'Fix login authentication error'"
You: "I need database access"

AI: "Sure! What database are you using?"

You: "PostgreSQL at localhost"

AI: "✓ PostgreSQL access configured!
     Restart your chat to start using database features."

Available integrations (15+): GitHub, PostgreSQL, Filesystem, Puppeteer, Brave Search, Slack, AWS, and more.

MCP Management Guide - Complete integration guide

How It Works

User: "I want to build an app"
  ↓
[Analyst Phase - Gathers requirements]
  ↓
[PM Phase - Creates development plan]
  ↓
[Architect Phase - Designs technical approach]
  ↓
[Scrum Master Phase - Breaks into stories]
  ↓
[Dev Phase - Guides implementation]
  ↓
[QA Phase - Validates quality]
  ↓
[All phases happen through natural conversation]

Dual-Lane Orchestration

Agilai automatically steers every request through the right mix of speed and depth. You never toggle a switch—the system senses complexity and elevates the workflow the moment your project needs more than a quick brief.

Quick Lane: 3-minute briefs

The quick lane boots up your docs/ workspace in minutes, generating the core folders and drafting the product requirement, architecture outline, user stories, and Google Nano Banana visual concept brief you need to get moving. It is designed for lightweight asks where momentum matters most, giving you production-ready assets without waiting on a full committee.

Complex Lane: BMAD depth on demand

When scope grows, the complex lane inherits everything the quick lane has already staged and then unleashes the full BMAD multi-agent orchestration. Analysts deepen the research, product managers expand the PRD, architects iterate on design, and delivery experts refine stories—all while building on the same documentation spine so nothing gets lost in translation.

Seamless hand-off, consistent outcomes

Because both lanes converge on the exact same deliverables, you can trust that every project—simple or sophisticated—arrives with the same professional PRD, architecture, and story set. The automatic hand-off means you stay in flow while Agilai keeps your docs/ assets synchronized and production-ready.

flowchart LR
    UserRequest["User request"] --> LaneSelector["Lane selector"]
    LaneSelector -->|Quick lane| QuickLane["Quick lane runtime"]
    LaneSelector -->|Complex lane| ComplexLane["Complex lane workflow"]
    QuickLane --> Docs["Shared docs/ outputs"]
    ComplexLane --> Docs

Summary: The Quick Lane runtime sequentially loads the Spec Kit–derived spec, plan, and tasks templates—calling the LLM at each step to author docs/prd.md, docs/architecture.md, and docs/stories/*.md—while the Complex lane employs the multi-agent BMAD workflow to produce the same deliverables through coordinated specialists.

  • Spec Kit Step – Product Spec TemplateBMAD Phase – Analyst Discovery & Product Framing
  • Spec Kit Step – Delivery Plan TemplateBMAD Phase – Architecture & Planning Alignment
  • Spec Kit Step – Tasks TemplateBMAD Phase – Story Breakdown & Delivery Execution

DUAL_LANE_ORCHESTRATION.md - Technical details

LLM Provider Support

Choose your provider during installation:

# Use ZhipuAI GLM
npx agilai@latest start --glm

# Use Anthropic Claude (default)
npx agilai@latest start --anthropic

Configuration Guide - Complete configuration reference

Documentation

Current Status

FULLY IMPLEMENTED AND PRODUCTION-READY

What Works

✅ MCP-based orchestration with 10 tools ✅ Natural conversation interface ✅ Intelligent dual-lane routing ✅ Phase detection and transitions ✅ Deliverable generation (PRD, architecture, stories) ✅ Project state persistence ✅ Full BMAD agent integration ✅ Claude, Codex, and OpenCode CLI support ✅ ZhipuAI GLM and Anthropic providers

Generated Artifacts

my-project/
├── docs/
│   ├── brief.md                        # Project context
│   ├── prd.md                          # Product requirements
│   ├── architecture.md                 # Technical design
│   ├── epics/*.md                      # Feature epics
│   ├── stories/*.md                    # User stories
│   └── ui/
│       ├── nano-banana-brief.md        # Google Nano Banana visual concept prompt
│       └── nano-banana-explorations.md # Logged visual concept selections
└── .agilai/
    ├── state.json                      # Current phase
    └── conversation.log                # Full history

Contributing

Contributions welcome! Key areas:

  • Additional test coverage
  • Custom template variations
  • Integration examples
  • Performance optimizations
  • Documentation improvements

See CONTRIBUTING.md for guidelines.

Related Projects

License

MIT License - see LICENSE for details.

Acknowledgments

Built on BMAD-METHOD™ by Brian (BMad) Madison.

BMAD™ and BMAD-METHOD™ are trademarks of BMad Code, LLC.


Current Version: v1.4.0 - Production Ready ✅

Get Started: npx agilai@latest start

Questions? Open an issue or check the BMAD repository.

Making AI-assisted development accessible to everyone through natural conversation