npm package discovery and stats viewer.

Discover Tips

  • General search

    [free text search, go nuts!]

  • Package details

    pkg:[package-name]

  • User packages

    @[username]

Sponsor

Optimize Toolset

I’ve always been into building performant and accessible sites, but lately I’ve been taking it extremely seriously. So much so that I’ve been building a tool to help me optimize and monitor the sites that I build to make sure that I’m making an attempt to offer the best experience to those who visit them. If you’re into performant, accessible and SEO friendly sites, you might like it too! You can check it out at Optimize Toolset.

About

Hi, 👋, I’m Ryan Hefner  and I built this site for me, and you! The goal of this site was to provide an easy way for me to check the stats on my npm packages, both for prioritizing issues and updates, and to give me a little kick in the pants to keep up on stuff.

As I was building it, I realized that I was actually using the tool to build the tool, and figured I might as well put this out there and hopefully others will find it to be a fast and useful way to search and browse npm packages as I have.

If you’re interested in other things I’m working on, follow me on Twitter or check out the open source projects I’ve been publishing on GitHub.

I am also working on a Twitter bot for this site to tweet the most popular, newest, random packages from npm. Please follow that account now and it will start sending out packages soon–ish.

Open Software & Tools

This site wouldn’t be possible without the immense generosity and tireless efforts from the people who make contributions to the world and share their work via open source initiatives. Thank you 🙏

© 2026 – Pkg Stats / Ryan Hefner

@pageai/ralph-loop

v1.21.0

Published

A long-running AI agent loop. Ralph automates software development tasks by iteratively working through a task list until completion.

Readme

A Ralph Wiggum Loop implementation that works™

@pageai/ralph-loop version

Ralph is a long-running AI agent loop. Ralph automates software development tasks by iteratively working through a task list until completion. This allows for long running agent loops, effectively enabling AI to code for days at a time.

This is an implementation that actually works, containing a hackable script so you can configure it to your env and favorite agentic AI CLI. It's set up by default to use Claude Code in a Docker sandbox, but supports many other agentic AI CLIs.

👉 Watch the video for an in-depth walkthrough.

Ralph Wiggum Loop


Getting Started

(Optional) Set up code base

I recommend using a CLI to bootstrap your project with the necessary tools and dependencies, e.g.:

npx @tanstack/cli create lib --add-ons eslint,form,tanstack-query,nitro --no-git

If you must start from a blank slate, which is not recommended, see Starting from scratch. You can also go for a more barebone start by running npx create-vite@latest src --template react-ts

1️⃣ Step 1: Install Ralph

Run this in your project's directory to install Ralph.

npx @pageai/ralph-loop

2️⃣ Step 2: Create a PRD + task list

Use the prd-creator skill to generate a PRD from your requirements. Open up Claude Code or Cursor etc. and prompt it with your requirements. Like so:

Use the prd-creator skill to help me create a PRD and task list for the below requirements.

An app is already set up with React, Tailwind CSS and TypeScript.

Requirements:

- A SaaS product that helps users manage their finances.
- Target audience: Small business owners and freelancers.
- Core features:
  - Track income and expenses.
  - Create and send invoices.
  - Track payments and receipts.
  - Generate reports and insights.
  - Connect to bank accounts and credit cards.
  - Connect to accounting software.
  - Connect to payment processors.
- Use the shadcn/ui library for components.
- Integrate with Stripe for payments.
- Use Supabase for database.
- You can find env variables in the .env.example file: SUPABASE_URL, SUPABASE_PUBLISHABLE_KEY, STRIPE_SECRET_KEY, etc. are available in the runtime.

// etc.

Check out the video for a more realistic example on how to write requirements.

  • mention libraries and frameworks you want to use
  • mention env variables, e.g. for DB, 3rd party API keys, etc. Store them in .env and add it to .gitignore
  • describe user flows and journeys
  • add relevant docs and UI references if applicable inside /docs and mention them in the requirements
  • be as descriptive as possible
  • it's fine to write in your own language

Then, follow the Skill's instructions and verify the PRD and then tasks. It is highly recommended that you review individual task requirements before starting the loop. Review EACH TASK INDIVIDUALLY.

3️⃣ Step 3: Set up the agent inside Docker sandbox

Authenticate inside the Docker sandbox before running Ralph. Run:

docker sandbox run claude .

And follow the instructions to log in into Claude Code.

👉 Answer "Yes" to Bypass Permissions mode, that's the exact reason why you are using the Docker sandbox.

If you want to use a different agentic CLI, see Running with a different agentic CLI.

4️⃣ Step 4: Run Ralph

./ralph.sh -n 50 # Run Ralph Loop with 50 iterations

✍️ Note: the first iteration will be spent on ensuring the sandbox environment is set up correctly. Expect 5 minutes to complete.

Running the Ralph Loop with custom options

# Run the agent loop (default: 10 iterations)
./ralph.sh

# Run with custom iteration limit
./ralph.sh 5
./ralph.sh -n 5
./ralph.sh --max-iterations 5

# Run exactly one iteration
./ralph.sh --once

# Show help
./ralph.sh --help

NB: you might need to run chmod +x ralph.sh to make the script executable.

The default "mode" is "implementation". Depending on your use case, you might want to change .agent/PROMPT.md to a different mode, e.g. "refactor", "review", "test" etc.

⚠️ If you want to use a different language or testing framework, see below.

(Optional) Adjusting to your language/framework

This script assumes the following are installed:

If you'd like to use a different language, testing framework etc. please adjust .agent/PROMPT.md to reflect your setup, server ports and startup commands etc.

👉 The loop is controlled by this prompt, which will be sent to the agent each iteration.


How It Works

Each iteration, Ralph will:

  1. Find the highest-priority incomplete task from .agent/tasks.json
  2. Work through the task steps defined in .agent/tasks/TASK-{ID}.json
  3. Run tests, linting, and type checking
  4. Complete task, take screenshot, update task status and commit changes
  5. Repeat until all tasks pass or max iterations reached

How Is This Different from Other Ralphs?

This was kept hackable so you can make it your own. The script follows the original concepts of the Ralph Wiggum Loop, working with fresh contexts and providing clear verifiable feedback.

It also works generically with any task set.

  • PRD generation - Creates a PRD and task list from requirements
  • Task lookup table generation - Creates a task lookup table from the PRD
  • Task breakdown + step generation - Breaks down each task into manageable steps
  • Iteration tracking - Shows progress through iterations with timing
  • Stream preview - Shows live output from the Agent
  • Step detection - Identifies current activity (Thinking, Implementing, Testing, etc.)
  • Screenshot capture - Captures a screenshot of the current screen
  • Notifications - Alerts when human input is needed
  • History logging - Saves clean output from each iteration
  • Timing - Shows timing metrics for each iteration and total time
  • Steering - Allows prioritizing critical work that needs to be done before the loop can continue
  • it allows you to dump unstructured requirements and have the agent create a PRD and task list for you.
  • it uses a task lookup table with individual detailed steps → more scalable as you get 100s of tasks done.
  • it's sandboxed and more secure
  • it shows progress and stats so you can keep an eye on what's been done
  • it instructs the agent to write and run automated tests and screenshots per task
  • it provides observability and traceability of the agent's work, showing a stream of output and capturing full historical logs per iteration

Steering the Agent

In some cases, you might notice the agent is having trouble, slowed down or struggling to overcome a blocker.

While the loop is running, you can edit the .agent/STEERING.md file to add critical work that needs to be done before the loop can continue.

The agent will check this file each iteration and if it finds any critical work, it will skip tasks and complete the critical work first.

Support

The ralph.sh script is designed to be hackable. It is configured to use Claude Code in a Docker sandbox by default, but with a one-liner change you can change it to use any other agentic AI CLI.

Check the ralph.sh script around # This is the main command loop. for the main command loop.

NB: skills are supported by all major agentic AI CLIs via symlinks.

Promise Tags

Ralph uses semantic tags to communicate status:

  • <promise>COMPLETE</promise> - All tasks finished successfully
  • <promise>BLOCKED:reason</promise> - Agent needs human help
  • <promise>DECIDE:question</promise> - Agent needs a decision

Exit Codes

| Code | Meaning | | ---- | ------------------------------ | | 0 | COMPLETE - All tasks finished | | 1 | MAX_ITERATIONS - Reached limit | | 2 | BLOCKED - Needs human help | | 3 | DECIDE - Needs human decision |

Structure

.agent/
├── PROMPT.md           # Prompt sent to Agent each iteration
├── tasks.json          # Task lookup table (required)
├── tasks/              # Individual task specs (TASK-{ID}.json)
├── prd/
│   ├── PRD.md          # Product requirements document
│   └── SUMMARY.md      # Short project overview sent to Agent each iteration
├── logs/
│   └── LOG.md          # Progress log (auto-created)
├── history/            # Iteration output logs
└── skills/             # Shared skills (source of truth)

Continued Development

As Ralph implements tasks, you might notice that you want some tweaks, features or even bug fixes. To do so, you need to continue using the prd-creator skill to update the PRD and task list.

For example:

I would like to expand the PRD. Use the prd-creator skill to create these tasks:

- there is a bug in X that should be fixed by doing Y
- create a new feature that implement Z

etc.

It's fine to add multiple tasks at once.

Skills

Skills are reusable agent capabilities that provide specialized knowledge and workflows. The canonical source is .agent/skills/, which is symlinked to multiple agent tool directories for compatibility.

Available Skills

| Skill | Description | | ----------------------------- | ------------------------------------------------------- | | component-refactoring | Patterns for splitting and refactoring React components | | e2e-tester | End-to-end testing workflows | | frontend-code-review | Code quality and performance review guidelines | | frontend-testing | Unit and integration testing patterns | | prd-creator | Create PRDs and task breakdowns for Ralph | | skill-creator | Create new skills | | vercel-react-best-practices | React/Next.js performance patterns | | mysql | MySQL/InnoDB schema, indexing, query tuning, and ops | | postgres | PostgreSQL best practices and query optimization | | web-design-guidelines | UI/UX design principles |

Skills Directory Structure

Skills are symlinked from .agent/skills/ to multiple locations for cross-tool compatibility:

 # Source of truth
.agent/skills/
    ├── component-refactoring/
    ├── e2e-tester/
    ├── postgres/
    ├── ...

# Symlinks -> .agent/skills/*
.agents/skills/*
.claude/skills/*
.codex/skills/*
.cursor/skills/*

Reference

Playwright configuration

If you are using Playwright, here is a recommended configuration:

import { defineConfig, devices } from '@playwright/test';

/**
 * See https://playwright.dev/docs/test-configuration.
 */
export default defineConfig({
  testDir: './tests',
  fullyParallel: true,
  globalTimeout: 30 * 60 * 1000,
  forbidOnly: !!process.env.CI,
  retries: process.env.CI ? 2 : 1,
  workers: process.env.CI ? 3 : 6,
  reporter: 'html',
  use: {
    baseURL: 'http://localhost:3000',
    trace: 'on-first-retry',
  },


  // NB: only chromium will run in Docker (arm64).
  projects: [
    {
      name: 'chromium',
      use: { ...devices['Desktop Chrome'] },
    }
  ],
});

Vitest configuration

If you are using Vitest, here is a recommended configuration:

import { defineConfig } from "vitest/config";
import react from "@vitejs/plugin-react";
import path from "path";

export default defineConfig({
  plugins: [react()],
  test: {
    environment: "node",
    globals: true,
    include: ["lib/**/*.test.ts", "lib/**/*.test.tsx"],
    // setupFiles: ['./vitest.setup.ts'], // Include this if using Next.js
  },
  resolve: {
    alias: {
      "@": path.resolve(__dirname),
    },
  },
});

If you are using Next.js, you'll also need a vitest.setup.ts file to mock the next/image and next/link components.

import '@testing-library/jest-dom/vitest'
import { vi } from 'vitest'
import React from 'react'

// If using Next.js, mock next/image
vi.mock('next/image', () => ({
  default: ({ src, alt, ...props }: { src: string; alt: string }) => {
    return React.createElement('img', { src, alt, ...props })
  },
}))

// If using Next.js, mock next/link
vi.mock('next/link', () => ({
  default: ({
    children,
    href,
    ...props
  }: {
    children: React.ReactNode
    href: string
  }) => {
    return React.createElement('a', { href, ...props }, children)
  },
}))

Running with a different agentic CLI

If you want to use a different agentic CLI, you can adjust the ralph.sh script to reflect your CLI of choice.

Check the ralph.sh script around # This is the main command loop. for the main command loop.

Replace docker sandbox run claude . -- with the your favorite CLI. Remember to also update the options after the --.

docker sandbox run codex . # for Codex CLI
docker sandbox run gemini . # for Gemini CLI

Docker currently supports: claude, codex, opencode,copilot, gemini, cagent, kiro and more. See all supported agentic AI CLIs in Docker's docs.

Starting from scratch

For AI to actually verify its implementation and for the loop to work, you need a way to verify it.

To that end, at the minimum you'll need an end-to-end test framework and a unit test framework.

For example, you can use the following commands to install Playwright and Vitest:

npm i @playwright/test vitest jsdom typescript eslint prettier -D

# If using React, also recommend installing:
npm i @vitejs/plugin-react @testing-library/dom @testing-library/jest-dom @testing-library/react @testing-library/user-event -D

It is recommended that you add skills for your specific language and framework. See skills.sh to discover existing skills.

Debugging

How to inspect the sandbox and debug

You might be wondering... if this is not a Docker container, how can you see what's going on inside Docker!? How to debug/install things?

That's quite straightforward.

You first need to run:

docker sandbox list

Note down the name of the sandbox, e.g. claude-ralph-loop.

And then you can run bash into any of the sandboxes like so:

docker sandbox exec -it <sandbox-name> bash # e.g. docker sandbox exec -it claude-ralph-loop bash

And you have full control over the sandbox, just like a regular container. You can install packages, run commands, etc.

You can also also run Claude Code inside the sandbox (make sure to navigate to the project directory first).

docker sandbox exec -it <sandbox-name> bash
cd /path/to/your/project # this is the same path as the path in the root machine, e.g. /Users/your-username/Documents/your-project
claude

License

MIT