npm package discovery and stats viewer.

Discover Tips

  • General search

    [free text search, go nuts!]

  • Package details

    pkg:[package-name]

  • User packages

    @[username]

Sponsor

Optimize Toolset

I’ve always been into building performant and accessible sites, but lately I’ve been taking it extremely seriously. So much so that I’ve been building a tool to help me optimize and monitor the sites that I build to make sure that I’m making an attempt to offer the best experience to those who visit them. If you’re into performant, accessible and SEO friendly sites, you might like it too! You can check it out at Optimize Toolset.

About

Hi, 👋, I’m Ryan Hefner  and I built this site for me, and you! The goal of this site was to provide an easy way for me to check the stats on my npm packages, both for prioritizing issues and updates, and to give me a little kick in the pants to keep up on stuff.

As I was building it, I realized that I was actually using the tool to build the tool, and figured I might as well put this out there and hopefully others will find it to be a fast and useful way to search and browse npm packages as I have.

If you’re interested in other things I’m working on, follow me on Twitter or check out the open source projects I’ve been publishing on GitHub.

I am also working on a Twitter bot for this site to tweet the most popular, newest, random packages from npm. Please follow that account now and it will start sending out packages soon–ish.

Open Software & Tools

This site wouldn’t be possible without the immense generosity and tireless efforts from the people who make contributions to the world and share their work via open source initiatives. Thank you 🙏

© 2025 – Pkg Stats / Ryan Hefner

@i-am-bee/beeai-supervisor

v0.0.13

Published

![cover](https://github.com/user-attachments/assets/8e61bb06-7548-4d57-9d00-897c6f4da414)

Readme

cover

Bee simplest circle BeeAI Supervisor

[!WARNING] [PRE-Alpha] This repository contains the beeai supervisor which is still under a rapid development. Please treat it as highly experimental and expect breaking changes often. Reach out on discord if you'd like to contribute or get involved in the discussions: join discord

Table of Contents

🧭 Overview

🌅 Quick Intro - beeai-supervisor.pdf 📚 Full documentation

BeeAI Supervisor is a multi-agent AI system designed to orchestrate specialized AI agents for completing complex tasks. While the name might suggest a single agent, it actually refers to the entire system - a comprehensive runtime that includes a Supervisor Agent at its core which serves as your primary interface, helping to configure, manage, and coordinate other specialized agents.

Think of BeeAI Supervisor as a hive of AI agents, with the Supervisor agent acting as the queen bee, orchestrating the activities of worker agents to efficiently accomplish your goals. This system enables you to harness the collective intelligence of multiple specialized agents through a single, coordinated interface.

🧑‍🏫 Getting started

🛠️ Installation

This project uses Mise-en-place as a manager of tool versions (python, uv, nodejs, pnpm etc.), as well as a task runner and environment manager. Mise will download all the needed tools automatically -- you don't need to install them yourself.

Clone this project, then run these setup steps:

brew install mise  # more ways to install: https://mise.jdx.dev/installing-mise.html
mise trust
mise install
mise build

🏡 Environment setup

Mise automatically creates a .env file based on the .env.template found in the project root. You must select one of the available LLM providers (ibm_rits, ollama, or openai) and set up its corresponding API key.

For example, if you choose OpenAI, your .env file might look like this:

# LLM Provider (ibm_rits/ollama/openai)
LLM_BACKEND="openai"

  :
  :

## OpenAI
OPENAI_API_KEY="<YOUR_OPEN_AI_API_KEY_HERE>"
OPENAI_MODEL_SUPERVISOR="gpt-4o"
OPENAI_MODEL_OPERATOR="gpt-4o"

[!WARNING] Ollama

If you choose the Ollama backend, ensure that all models you plan to use support tool calling. Using a small model can lead to many incorrect tool calls from the supervisor agent. For stable performance, we recommend a large model such as qwq:32b.

🎛️ Operation Modes

The system can be operated in two distinct modes: interactive and autonomous.

💬 Interactive Mode

chat

In this mode, you engage in an ongoing conversation with the Supervisor in Chat UI:

mise interactive

This command starts the platform in interactive mode where you can:

  • Start a Conversation: Interact with the Supervisor Agent to describe your goals
  • Iterative Configuration: Collaborate with the Supervisor to define appropriate agent and task configurations
  • Refine Your Approach: Based on results, work with the Supervisor to improve your setup
  • Adjust On the Fly: You can abort an ongoing task and refine the assignment through chat with the Supervisor

🤖 Autonomous Mode

For one-off tasks or batch processing, you can run the platform in autonomous mode:

mise autonomous <<< "Hi, can you create a poem about each of these topics: bee, hive, queen, sun, flowers?"

This command runs the platform in autonomous mode where you can:

  • One-Shot Requests: Provide your instruction or data in a single command
  • Hands-Off Processing: The Supervisor Agent autonomously orchestrates the required agents and tasks
  • Focused Efficiency: No ongoing conversation is required, letting the system complete your task in the background
  • Automatic Wrap-Up: Once the request is fulfilled, results are returned and the platform shuts down

Autonomous mode is ideal for scripted operations, batch processing, or one-off requests that don’t require iterative development. Once your request is complete, results are returned, and the system automatically shuts down—no further input or interaction needed.

🖥️ Monitor

monitors

To monitor the process while in both modes, open another terminal window and run:

mise monitor

This allows you to observe the platform's operations as it works on your request.

📤 Workspaces

One of the main challenges in a multi-agent system is preventing the creation of unnecessary objects that can quickly consume resources. To address this, Workspaces provide a persistence layer for both agent and task configurations. This layer preserves configuration states across sessions and supports iterative refinement. Once you’ve fine-tuned a configuration to meet your needs, you can reuse it for new inputs—avoiding the expense of rebuilding complex structures, ensuring consistent processing quality, and optimizing token usage.

Workspaces are located in ./outputs/workspaces folder:

workspaces

You can create or switch a workspace by setting an environment variable like this:

WORKSPACE=my_workspace mise interactive

🎪 Showcase

🎶 Creative Multi-Agent Tasks Showcase: Poetry and Hip-Hop Analysis

The Poetry and Hip-Hop Analysis showcase demonstrates BeeAI Supervisor's sophisticated task orchestration capabilities. The system efficiently manages complex task dependencies, enabling parallel execution of multiple poem and song analyses while maintaining synchronization points where needed. Multiple agent instances work simultaneously on different aspects of the analysis, with the supervisor ensuring proper task sequencing and resource allocation.

This showcase particularly highlights the system's ability to handle both parallel and sequential task flows, dynamically adjusting agent allocation based on task dependencies and availability. The supervisor maintains coherent coordination across multiple concurrent analysis streams while preventing resource conflicts and ensuring optimal throughput.

⚙️ Specification

  • 🤹‍♂️ Interactive, 🤖 Autonomous, 🖼️ No-code
  • Used agents
    • boss 1x
      meta-llama/llama-3-1-405b-instruct-fp
    • peom_generator 4x meta-llama/llama-3-3-70b-instruct
    • hip_hop_song_generator 1x meta-llama/llama-3-3-70b-instruct
    • poem_elements_highlighter 1x meta-llama/llama-3-3-70b-instruct
  • Executed tasks
    • process_input_and_plan 1x
    • poem_generation 4x
    • hip_hop_song_generation 1x
    • poem_elements_highlighting 1x

🏁 Conclusion

This platform combines the power of specialized AI agents with the flexibility of interactive configuration, all managed through natural conversation with the Supervisor Agent.

🚀 Key Benefits

  • Iterative Development: Refine agents and tasks through ongoing dialogue with the Supervisor, ensuring continuous improvement.
  • Workspace Persistence: Preserve and reuse configurations to save tokens and maintain consistent results across sessions.
  • Parallel Scalability: Coordinate specialized agents to handle complex or high-volume operations efficiently.
  • Unified Interface: Control all AI agents from a single entry point, simplifying orchestration.
  • Active Monitoring: Gain real-time insights into platform operations to quickly identify and address bottlenecks.

💡 Best Practices

  • Incremental Approach: Start with smaller tasks and gradually expand to more complex, multi-agent workflows.
  • Leverage Workspaces: Reuse proven agent and task configurations to optimize resource usage and maintain quality.
  • Stay Observant: Use the Monitor to track operations and make timely adjustments based on system feedback.
  • Document Configurations: Keep a record of effective setups and methods for easier knowledge sharing and future reference.