npm package discovery and stats viewer.

Discover Tips

  • General search

    [free text search, go nuts!]

  • Package details

    pkg:[package-name]

  • User packages

    @[username]

Sponsor

Optimize Toolset

I’ve always been into building performant and accessible sites, but lately I’ve been taking it extremely seriously. So much so that I’ve been building a tool to help me optimize and monitor the sites that I build to make sure that I’m making an attempt to offer the best experience to those who visit them. If you’re into performant, accessible and SEO friendly sites, you might like it too! You can check it out at Optimize Toolset.

About

Hi, 👋, I’m Ryan Hefner  and I built this site for me, and you! The goal of this site was to provide an easy way for me to check the stats on my npm packages, both for prioritizing issues and updates, and to give me a little kick in the pants to keep up on stuff.

As I was building it, I realized that I was actually using the tool to build the tool, and figured I might as well put this out there and hopefully others will find it to be a fast and useful way to search and browse npm packages as I have.

If you’re interested in other things I’m working on, follow me on Twitter or check out the open source projects I’ve been publishing on GitHub.

I am also working on a Twitter bot for this site to tweet the most popular, newest, random packages from npm. Please follow that account now and it will start sending out packages soon–ish.

Open Software & Tools

This site wouldn’t be possible without the immense generosity and tireless efforts from the people who make contributions to the world and share their work via open source initiatives. Thank you 🙏

© 2026 – Pkg Stats / Ryan Hefner

openclown

v0.2.2

Published

Multi-perspective AI task evaluation circus for OpenClaw

Readme

🎪 OpenClown — Multi-Perspective AI Task Evaluation

OpenClown is a plugin for OpenClaw that evaluates AI-completed tasks from multiple expert perspectives. A "circus" of specialized performers — a philosopher, a security expert, a developer, and more — independently critique what your AI assistant just did.

The goal: catch blind spots, surface risks, and improve AI output quality before you act on it.

How it works

You ask OpenClaw to do something
        │
        ▼
  OpenClaw completes the task
        │
        ▼
  You reply to the response with /clown
  (or just type /clown for the latest)
        │
        ▼
┌─────────────────────────────────────┐
│         OpenClown Circus            │
│                                     │
│  📎 Identify target exchange        │
│     (reply content matching,        │
│      keyword, or latest)            │
│  📝 Gather context                  │
│     (target + up to 3 prior         │
│      exchanges for follow-ups       │
│      + tool calls & results)        │
│              │                      │
│  🎭 Philosopher  → questions assumptions
│  🔒 Security     → flags data exposure
│  💻 Developer    → spots better approaches
│  ... (12 performers available)      │
└──────────────┬──────────────────────┘
               │
               ▼
     Formatted evaluation with
     severity levels + actionable feedback

Install

openclaw plugins install openclown

That's it. OpenClown automatically reuses the LLM provider and API key you configured during OpenClaw setup — no extra configuration needed.

Update to latest version:

openclaw plugins update openclown

Quick start

On mobile (WhatsApp, Telegram, Slack, Discord):

  1. Ask OpenClaw a question and wait for the response
  2. Long-press (or swipe) the AI response you want to evaluate
  3. Tap Reply
  4. Type /clown and send

That's it. OpenClown identifies the message you replied to and evaluates it with full context — including the original question, any tool calls the AI made, and prior conversation history.

If you don't reply to a specific message, /clown automatically evaluates the most recent AI response.

| Channel | Reply + /clown | /clown (latest) | /clown <keyword> | |---------|:-:|:-:|:-:| | WhatsApp | ✅ | ✅ | ✅ | | Telegram | ✅ | ✅ | ✅ | | Discord | ✅ | ✅ | ✅ | | Slack | ✅ | ✅ | ✅ | | CLI | — | ✅ | ✅ |

/clown — Evaluate

Reply to any AI response and type /clown to evaluate it. Or just send /clown to evaluate the latest response.

/clown                  # Evaluate latest AI response
/clown <keyword>        # Search by keyword, e.g. /clown weather

Each evaluation shows results from multiple performers with severity levels:

  • 💡 Insight — observations, suggestions, things to consider
  • ⚠️ Warning — potential risks or issues worth addressing
  • 🔴 Critical — serious problems that should be fixed

/clown encore — Improve

After an evaluation, reply with /clown encore to re-run the original task with the evaluation feedback applied. The AI re-answers your question, addressing the issues the performers found:

/clown encore           # AI re-answers, addressing the issues found

Usage examples

OpenClown works anywhere OpenClaw does — WhatsApp, Telegram, Slack, Discord, or CLI.

Ask a question, then evaluate

/clown — multi-perspective evaluation

/clown encore — improved response with feedback applied

Reply targeting

Reply to any AI response — not just the latest — and type /clown to evaluate that specific exchange:

[Reply to any OpenClaw message] /clown
→ 📎 Matched reply: "Yes! Toronto public pools are open on Sunday..."
→ Evaluates that specific response with full context

Follow-up questions

OpenClown automatically includes up to 3 prior exchanges as context, so evaluators understand follow-up questions:

You:     Which cities in Canada are best for startups?
OpenClaw: Vancouver and Toronto lead the pack...

You:     How about Ottawa?
OpenClaw: Ottawa has a growing tech scene...

You:     /clown
         → evaluators see both exchanges and understand
           "How about Ottawa?" means "Is Ottawa good for startups?"

Performers

OpenClown ships with 12 performers. Three are enabled by default:

| Performer | Emoji | Focus | |-----------|-------|-------| | Philosopher | 🎭 | Assumptions, definitions, epistemic honesty | | Security Expert | 🔒 | Data exposure, API security, privacy | | Developer | 💻 | Implementation quality, error handling, efficiency |

Additional performers you can enable:

| Performer | Emoji | Focus | |-----------|-------|-------| | Ethicist | ⚖️ | Fairness, inclusivity, potential harm | | Fact Checker | 🔍 | Accuracy, sources, hallucination detection | | UX Designer | 🎨 | Information hierarchy, scannability, actionability | | VC Investor | 💰 | Value proposition, scalability, ROI | | Comedian | 😂 | Absurdity, overthinking, unintentional humor | | Shakespeare | 🎭 | Narrative arc, emotional truth, prose quality | | Conspiracy Theorist | 🔮 | Data provenance, hidden agendas, algorithmic bias | | Grandparents | 👵👴 | Practicality, common sense, well-being | | Cat Expert | 🐱 | Efficiency, priorities, power dynamics |

Managing your circus

Use /clown circus to see your current lineup. Enable or disable performers using their id or number:

/clown circus on comedian       # Enable by id
/clown circus on 4,5            # Enable by number
/clown circus off philosopher   # Disable by id
/clown circus off 1             # Disable by number
/clown circus toggle 1,3,8      # Switch multiple on/off in one command
/clown circus reset             # Restore defaults (Philosopher, Security, Developer)

Your lineup is saved to ~/.openclaw/openclown/circus.json and persists across restarts.

Create your own performer

Don't see the perspective you need? Create your own through a guided conversational flow — no code or config files needed:

You:     /clown circus create A maritime law expert who evaluates
         responses for legal accuracy around shipping regulations

OpenClaw: 🎪 Creating a new performer...

         1. What specific aspects of maritime law should this
            evaluator focus on?
         2. Should the evaluation style be formal/checklist-based
            or more conversational?
         3. How severe should findings be — advisory insights,
            warnings, or critical errors?
         ...

You:     /clown circus create Focus on UNCLOS compliance and
         cargo liability. Formal style. Severity: warning.

OpenClaw: 🎪 Here's your performer draft:
         ━━━━━━━━━━━━━━━━━━━━━━
         ⚖️ Maritime Law Expert [maritime]
         Severity: ⚠️ Warning
         Category: serious
         ━━━━━━━━━━━━━━━━━━━━━━

         /clown circus confirm — save and enable
         /clown circus preview — see full definition
         /clown circus create <changes> — revise
         /clown circus cancel — discard

You:     /clown circus confirm

OpenClaw: 🎪 New Performer Created!
         ⚖️ Maritime Law Expert [maritime]
         ✅ Saved and enabled.

Custom performers are saved to ~/.openclaw/openclown/skills/ and work exactly like built-in ones. You can also edit or delete them:

/clown circus edit maritime Make it focus more on cargo liability
/clown circus delete maritime

For the full command reference (configuration, manual SKILL.md format, all options), see the Command Reference.

Multilingual support

OpenClown auto-detects the language of the user's request and evaluates in the same language. Supported locales: English, Chinese, Japanese, Korean, French, Spanish.

Development

git clone https://github.com/Bubbletea98/openclown.git
cd openclown
npm install
npm run build
npm test

See CONTRIBUTING.md for contribution guidelines.

License

MIT