npm package discovery and stats viewer.

Discover Tips

  • General search

    [free text search, go nuts!]

  • Package details

    pkg:[package-name]

  • User packages

    @[username]

Sponsor

Optimize Toolset

I’ve always been into building performant and accessible sites, but lately I’ve been taking it extremely seriously. So much so that I’ve been building a tool to help me optimize and monitor the sites that I build to make sure that I’m making an attempt to offer the best experience to those who visit them. If you’re into performant, accessible and SEO friendly sites, you might like it too! You can check it out at Optimize Toolset.

About

Hi, 👋, I’m Ryan Hefner  and I built this site for me, and you! The goal of this site was to provide an easy way for me to check the stats on my npm packages, both for prioritizing issues and updates, and to give me a little kick in the pants to keep up on stuff.

As I was building it, I realized that I was actually using the tool to build the tool, and figured I might as well put this out there and hopefully others will find it to be a fast and useful way to search and browse npm packages as I have.

If you’re interested in other things I’m working on, follow me on Twitter or check out the open source projects I’ve been publishing on GitHub.

I am also working on a Twitter bot for this site to tweet the most popular, newest, random packages from npm. Please follow that account now and it will start sending out packages soon–ish.

Open Software & Tools

This site wouldn’t be possible without the immense generosity and tireless efforts from the people who make contributions to the world and share their work via open source initiatives. Thank you 🙏

© 2026 – Pkg Stats / Ryan Hefner

@ckelsoe/claude-skill-prompt-architect

v2.1.4

Published

Claude Code skill for analyzing and improving prompts using 27 research-backed frameworks across 7 intent categories. Intent-based selection system routes to the right framework — from ultra-minimal (APE) to agentic (ReAct), with create, transform, reason

Readme

Claude Code Skill: Prompt Architect

Transform vague prompts into expert-level, structured prompts using 27 research-backed frameworks across 7 intent categories.

A comprehensive skill that analyzes, architects, and iteratively refines prompts through systematic framework application and guided dialogue.

License: MIT


Table of Contents


Overview

Prompt Architect is a production-ready Claude Code Skill that elevates your prompting capabilities through:

  • Intelligent Analysis - Evaluates prompts across 5 quality dimensions (clarity, specificity, context, completeness, structure)
  • Framework Recommendation - Suggests the best framework(s) for your specific use case with clear reasoning
  • Guided Dialogue - Asks targeted clarifying questions to gather missing information progressively
  • Systematic Application - Applies selected framework to transform your prompt
  • Iterative Refinement - Continues improving based on feedback until perfect

Target Audience:

  • Developers using Claude Code for software engineering tasks
  • Prompt engineers optimizing LLM interactions
  • AI practitioners seeking systematic prompt improvement
  • Teams wanting consistent, high-quality prompts

Key Features

27 Research-Backed Frameworks Across 7 Intent Categories

| Framework | Best For | Complexity | |-----------|----------|------------| | CO-STAR | Content creation, writing tasks | High | | RISEN | Multi-step processes, procedures | High | | CRISPE | Comprehensive prompts with multiple output variants | High | | BROKE | Business deliverables with OKR-style measurable outcomes | Medium | | RISE-IE | Data analysis, transformations (Input-Expectation) | Medium | | RISE-IX | Content creation with examples (Instructions-Examples) | Medium | | TIDD-EC | High-precision tasks with explicit dos/don'ts | Medium | | RACE | Expert tasks requiring role + context + outcome clarity | Medium | | CARE | Constraint-driven tasks with explicit rules and examples | Medium | | CTF | Simple tasks where situational context drives the prompt | Low | | RTF | Simple, focused tasks where expertise framing matters | Low | | APE | Ultra-minimal one-off prompts | Low | | BAB | Rewriting, refactoring, transforming existing content | Low | | Tree of Thought | Decisions requiring exploration of multiple approaches | Medium | | ReAct | Agentic / tool-use tasks with iterative reasoning | Medium | | Skeleton of Thought | Structured long-form content (outline-first) | Medium | | Step-Back | Principle-grounded reasoning (abstract first, then specific) | Medium | | Least-to-Most | Compositional multi-hop problems (simplest first) | Medium | | Plan-and-Solve (PS+) | Zero-shot numerical/calculation reasoning | Low | | Chain of Thought | Reasoning, problem-solving | Medium | | Chain of Density | Iterative refinement, summarization | Medium | | Self-Refine | Iterative output quality improvement (any task) | Medium | | CAI Critique-Revise | Principle-based critique and revision (Anthropic) | Medium | | Devil's Advocate | Strongest opposing argument against a position | Low | | Pre-Mortem | Assume failure, identify specific causes | Low | | RCoT | Verify reasoning by reconstructing the question | Medium | | RPEF | Recover/reconstruct a prompt from an existing output | Low | | Reverse Role Prompting | AI interviews you before executing | Low |

Quality Scoring System

Every prompt is evaluated across:

  • Clarity (1-10): Is the goal clear and unambiguous?
  • Specificity (1-10): Are requirements detailed enough?
  • Context (1-10): Is background information provided?
  • Completeness (1-10): Are all necessary elements present?
  • Structure (1-10): Is the prompt well-organized?

Progressive Disclosure

  • Asks 3-5 targeted questions at a time (never overwhelming)
  • Adapts questions based on framework selection
  • Builds understanding iteratively
  • Confirms assumptions before proceeding

Continuous Iteration

  • Shows before/after comparisons with explanations
  • Refines based on your feedback
  • Can switch frameworks if initial choice doesn't fit
  • Continues until you're satisfied

Example Transformation

Before

"Write about machine learning"

Analysis Scores:

  • Clarity: 2/10 (vague goal)
  • Specificity: 1/10 (no details)
  • Context: 0/10 (no background)
  • Completeness: 2/10 (missing most elements)
  • Structure: 3/10 (single sentence)
  • Overall: 1.6/10

After (CO-STAR Framework)

CONTEXT:
Creating content for a business blog aimed at C-level executives exploring
how AI/ML could benefit their organizations. Readers understand business
strategy but have limited technical ML knowledge. Part of an emerging
technologies series.

OBJECTIVE:
Create an engaging article helping executives understand practical machine
learning applications relevant to their companies. Focus on demonstrating
tangible business value and real-world implementation without overwhelming
technical details.

STYLE:
Professional blog style combining narrative with bullet points. Include 2-3
real-world case studies. Structure with clear subheadings every 150-200 words.
Balance storytelling with concrete information. Avoid jargon; when necessary,
provide plain-language explanations.

TONE:
Professional yet approachable and conversational. Confident and authoritative
without being condescending. Practical and business-focused rather than
theoretical.

AUDIENCE:
C-suite executives and senior managers at mid-to-large enterprises who:
- Make strategic technology investment decisions
- Understand business metrics and ROI
- Have limited technical ML knowledge
- Value practical examples over theory

RESPONSE FORMAT:
800-word article structured as:
- Compelling headline (10 words max)
- Brief hook (2-3 sentences)
- 3-4 main sections with descriptive subheadings
- Mix of paragraphs and bullet points
- Clear call-to-action conclusion

Result Scores:

  • Clarity: 9/10
  • Specificity: 9/10
  • Context: 10/10
  • Completeness: 9/10
  • Structure: 9/10
  • Overall: 8.8/10

Supported Frameworks

CO-STAR (Context, Objective, Style, Tone, Audience, Response)

Best for: Content creation, writing tasks, communications

Components:

  • Context - Background information and situation
  • Objective - Clear goal and purpose
  • Style - Writing style and formatting approach
  • Tone - Voice and emotional quality
  • Audience - Target reader characteristics
  • Response - Expected format and structure

Example Use Cases: Blog posts, emails, presentations, marketing copy, documentation


RISEN (Role, Instructions, Steps, End goal, Narrowing)

Best for: Multi-step processes, systematic procedures

Components:

  • Role - Expertise and perspective needed
  • Instructions - High-level guidance
  • Steps - Detailed methodology
  • End goal - Success criteria
  • Narrowing - Constraints and boundaries

Example Use Cases: Code reviews, workflows, systematic analysis, project planning


RISE (Dual Variants)

RISE-IE (Input-Expectation)

Best for: Data analysis, transformations, processing tasks

Components:

  • Role - Expertise needed
  • Input - Data format and characteristics
  • Steps - Processing methodology
  • Expectation - Output requirements

Example Use Cases: CSV analysis, data processing, file transformations, report generation

RISE-IX (Instructions-Examples)

Best for: Content creation with reference examples

Components:

  • Role - Expertise needed
  • Instructions - Task guidance
  • Steps - Workflow process
  • Examples - Reference samples

Example Use Cases: Creative writing, template-based content, style matching


TIDD-EC (Task, Instructions, Do, Don't, Examples, Context)

Best for: High-precision tasks requiring explicit boundaries

Components:

  • Task type - Nature of the work
  • Instructions - What to accomplish
  • Do - Explicit positive guidance
  • Don't - Explicit negative guidance (what to avoid)
  • Examples - Reference samples
  • Context - Background information

Example Use Cases: Code generation with standards, compliance tasks, quality-critical work


CTF (Context, Task, Format)

Best for: Simple tasks where situational background matters more than expertise framing

Components:

  • Context - Situation and background
  • Task - What needs to be done
  • Format - Output structure

Example Use Cases: Handoff documents, mid-project updates, situation-driven requests


RTF (Role, Task, Format)

Best for: Simple, well-defined tasks where expertise framing drives output quality

Components:

  • Role - Expertise required
  • Task - What needs to be done
  • Format - Output structure

Example Use Cases: Quick conversions, simple formatting, straightforward requests


APE (Action, Purpose, Expectation)

Best for: Ultra-minimal prompts — the simplest structured framework

Components:

  • Action - What to do (one clear verb-driven instruction)
  • Purpose - Why it's needed (one sentence)
  • Expectation - What a good result looks like

Example Use Cases: Quick summaries, single-function code, one-off requests, rapid iteration


BAB (Before, After, Bridge)

Best for: Transforming, rewriting, or refactoring existing content

Components:

  • Before - Current state and what's wrong
  • After - Desired end state
  • Bridge - Transformation rules and constraints

Example Use Cases: Code refactoring, copy rewrites, tone changes, document restructuring, version migrations


RACE (Role, Action, Context, Expectation)

Best for: Medium-complexity tasks needing expertise + background + explicit success criteria

Components:

  • Role - Expertise or persona
  • Action - What needs to be done
  • Context - Situational background and constraints
  • Expectation - What a good result looks like

Example Use Cases: Technical reviews, expert analysis, contextual recommendations, documentation with standards


CRISPE (Capacity+Role, Insight, Instructions, Personality, Experiment)

Best for: Comprehensive prompts where you want multiple output variants to compare

Components:

  • Capacity & Role - Expertise level and professional persona
  • Insight - Background context and situational knowledge
  • Instructions - The specific task
  • Personality - Tone, voice, communication style
  • Experiment - Request N variants to compare (the defining differentiator)

Example Use Cases: Marketing campaigns (A/B variants), content with tone options, strategic analysis needing multiple angles


BROKE (Background, Role, Objective, Key Results, Evolve)

Best for: Business deliverables with measurable outcomes and built-in self-improvement

Components:

  • Background - Situation and context
  • Role - Professional persona
  • Objective - The task
  • Key Results - Measurable business outcomes (OKR-style)
  • Evolve - AI self-critiques and suggests 3 improvements

Example Use Cases: Sales process improvements, content strategy with KPIs, product decisions tied to metrics


CARE (Context, Ask, Rules, Examples)

Best for: Tasks with explicit constraints, compliance requirements, or quality standards

Components:

  • Context - Situation and background
  • Ask - The specific request
  • Rules - Explicit constraints, dos/don'ts, standards
  • Examples - Reference samples showing the quality bar

Example Use Cases: Healthcare/legal content, UI error messages, interview questions with bias constraints, brand-compliant copy


Skeleton of Thought (SoT)

Best for: Structured long-form content — generate outline first, then expand

Approach:

  • Phase 1: Generate a concise skeleton/outline (key points only)
  • Phase 2: Expand each skeleton point independently (parallelizable)
  • SoT-R variant: Route non-suitable questions to standard generation

Example Use Cases: Technical documentation, structured reports, tutorials, any multi-section content


Step-Back Prompting

Best for: Principle-grounded reasoning — abstract to the underlying concept first

Approach:

  • Generate a higher-level "step-back" question about underlying principles
  • Answer the abstract question to retrieve principles
  • Use those principles as context to answer the original specific question

Example Use Cases: STEM problems, architecture decisions, debugging, any task where first-principles reasoning matters


Least-to-Most (LtM)

Best for: Compositional multi-hop problems with ordered dependencies

Approach:

  • Decompose the complex problem into ordered subproblems (simplest first)
  • Solve each subproblem sequentially, feeding each answer into the next
  • Use accumulated answers to solve the original problem

Example Use Cases: Multi-domain questions (legal + technical), complex calculations, architecture problems with prerequisites


Plan-and-Solve (PS+)

Best for: Zero-shot numerical and calculation reasoning

Approach:

  • "Let's first understand the problem, extract relevant variables and their corresponding numerals, and devise a complete plan. Then, let's carry out the plan, calculate intermediate values, pay attention to computation, and solve the problem step by step."

Example Use Cases: Financial calculations (MRR, CAC, payback), math word problems, resource estimation, any zero-shot reasoning task


Self-Refine

Best for: Iterative quality improvement of any output

Approach: Generate → produce specific actionable feedback → refine → repeat until stopping criterion Research: Madaan et al. NeurIPS 2023 — +5-40% improvement across 7 task types

Example Use Cases: Code review and rewriting, writing improvement, plan refinement, pre-submission QA


CAI Critique-Revise

Best for: Aligning output to an explicit stated principle or standard

Approach: Initial output → critique against a specific principle → revision addressing every critique point Research: Anthropic Constitutional AI (arXiv 2212.08073, 2022) — principle-driven alignment

Example Use Cases: Plain language compliance, brand voice enforcement, epistemic quality (claims vs. assertions), legal/regulatory language standards


Devil's Advocate

Best for: Generating the strongest possible opposing argument against a position

Approach: Explicitly instructs the AI to attack a position as forcefully as possible — not balanced, not a straw man, but maximum-strength opposition Research: ACM IUI 2024 peer-reviewed study

Example Use Cases: Decision stress-testing, architecture reviews, countering groupthink, stakeholder preparation, debiasing


Pre-Mortem

Best for: Identifying specific failure causes before they happen

Approach: Assume the project has already failed → describe the failure → work backwards to specific causes with warning signs Research: Gary Klein's prospective hindsight — ~30% improvement over forward risk analysis

Example Use Cases: Project kickoffs, product launches, technical migrations, high-stakes strategic decisions


RCoT (Reverse Chain-of-Thought)

Best for: Verifying that reasoning addressed all conditions in a multi-constraint question

Approach: Generate answer → reconstruct the question from the answer → cross-check conditions → correct overlooked items Research: Academic backward reasoning literature (ACL 2025, NAACL 2025)

Example Use Cases: Multi-condition logic problems, complex requirements analysis, high-stakes reasoning verification


RPEF (Reverse Prompt Engineering)

Best for: Recovering a reusable prompt template from an existing output

Approach: Provide an output (and optionally the input) → AI analyzes it for tone, structure, constraints, persona → generates a reusable template with [PLACEHOLDER] variables Research: Li & Klabjan, EMNLP 2025 (arXiv 2411.06729)

Example Use Cases: Recovering lost prompts, codifying successful one-time outputs, building style templates, understanding system prompt behavior


Reverse Role Prompting (AI-Led Interview)

Best for: Complex tasks where you know the goal but struggle to specify all requirements

Approach: Provide a minimal intent statement → AI asks targeted clarifying questions → executes once context is complete Research: FATA framework (arXiv 2508.08308, 2025) — ~40% improvement over standard prompting

Example Use Cases: Complex strategy tasks, non-expert users, requirements gathering, generating complete prompts from an interview


Tree of Thought

Best for: Decisions where multiple approaches need systematic comparison

Approach:

  • Defines the problem and constraints
  • Explores 2-5 distinct solution branches in parallel
  • Evaluates each branch against defined criteria
  • Synthesizes into a reasoned recommendation

Example Use Cases: Architecture decisions, debugging with multiple hypotheses, technology selection, strategic trade-offs


ReAct (Reasoning + Acting)

Best for: Agentic tasks that interleave reasoning with tool use

Approach:

  • Defines goal, available tools, and constraints
  • Alternates Thought → Action → Observation cycles
  • Each observation informs the next thought
  • Continues until goal is reached

Example Use Cases: Agentic workflows, multi-step research, debugging with tools, data investigation


Chain of Thought

Best for: Complex reasoning and problem-solving

Approach:

  • Breaks down reasoning into explicit steps
  • Shows work and intermediate conclusions
  • Verifies logic at each stage
  • Builds to final answer

Example Use Cases: Math problems, debugging, decision analysis, logical reasoning


Chain of Density

Best for: Iterative refinement and compression

Approach:

  • Starts with baseline version
  • Progressively refines through iterations
  • Increases information density
  • Optimizes for specific goals

Example Use Cases: Summarization, content compression, explanation optimization


Use with Other AI Tools

Prompt Architect works with Claude Code out of the box. For other tools, use the adapter files in the adapters/ folder.

| Tool | Adapter file | Where to put it | |------|-------------|-----------------| | OpenAI Codex CLI | adapters/for-openai-codex-cli.md | Append to AGENTS.md in project root | | Cursor | adapters/for-cursor.mdc | Copy to .cursor/rules/prompt-architect.mdc | | GitHub Copilot | adapters/for-github-copilot.md | Append to .github/copilot-instructions.md | | Windsurf | adapters/for-windsurf.md | Append to .windsurfrules in project root | | Any LLM API / ChatGPT | adapters/system-prompt.md | Paste into system prompt / instructions field |

Quick install examples:

# OpenAI Codex CLI
cat node_modules/@ckelsoe/claude-skill-prompt-architect/adapters/for-openai-codex-cli.md >> AGENTS.md

# Cursor
mkdir -p .cursor/rules
cp node_modules/@ckelsoe/claude-skill-prompt-architect/adapters/for-cursor.mdc .cursor/rules/prompt-architect.mdc

# GitHub Copilot
mkdir -p .github
cat node_modules/@ckelsoe/claude-skill-prompt-architect/adapters/for-github-copilot.md >> .github/copilot-instructions.md

# Windsurf
cat node_modules/@ckelsoe/claude-skill-prompt-architect/adapters/for-windsurf.md >> .windsurfrules

See adapters/README.md for full documentation, limitations, and troubleshooting for each platform.


Quick Start

1. Install via npm (Recommended)

# Install globally
npm install -g @ckelsoe/claude-skill-prompt-architect

# Run installation script
prompt-architect-install

Alternative: Manual Installation

If you prefer not to use npm:

macOS/Linux:

cp -r prompt-architect ~/.claude/skills/

Windows (PowerShell):

Copy-Item -Path "prompt-architect" -Destination "$env:USERPROFILE\.claude\skills\prompt-architect" -Recurse

2. Restart Claude Code

Close and reopen Claude Code to load the skill.

3. Try It Out

"Help me improve this prompt: write a technical blog post"

The skill will automatically:

  1. Analyze your prompt and score it
  2. Recommend the best framework (likely CO-STAR)
  3. Ask clarifying questions
  4. Generate an improved, structured prompt
  5. Explain the changes made

Installation

Method 1: npm (Recommended)

Option A: Global Installation (User-Wide)

# Install globally
npm install -g @ckelsoe/claude-skill-prompt-architect

# Run installation script
prompt-architect-install

# Restart Claude Code

Option B: Local Installation (Without Global)

# Install package
npm install @ckelsoe/claude-skill-prompt-architect

# Run installation script using npx
npx prompt-architect-install

# Restart Claude Code

Option C: Project-Specific Installation

# Install as dev dependency
npm install --save-dev @ckelsoe/claude-skill-prompt-architect

# Install to project .claude/skills/ folder
npx prompt-architect-install --project

# Restart Claude Code

Method 2: Manual Installation

If you prefer not to use npm:

Prerequisites

  • Claude Code installed and configured
  • Python 3.7+ (optional, for analysis scripts)

Installation Steps

  1. Clone or Download this Repository
git clone https://github.com/ckelsoe/claude-skill-prompt-architect.git
cd claude-skill-prompt-architect
  1. Copy Skill to Claude Code Directory

The skill needs to be in your Claude Code skills folder.

Default Locations:

  • Windows: C:\Users\<YourName>\.claude\skills\
  • macOS: ~/.claude/skills/
  • Linux: ~/.claude/skills/

Installation Commands:

# macOS/Linux
cp -r prompt-architect ~/.claude/skills/

# Windows (PowerShell)
Copy-Item -Path "prompt-architect" -Destination "$env:USERPROFILE\.claude\skills\prompt-architect" -Recurse

# Windows (Command Prompt)
xcopy prompt-architect %USERPROFILE%\.claude\skills\prompt-architect /E /I
  1. Verify Installation

Check that all files are present:

# macOS/Linux
ls ~/.claude/skills/prompt-architect/

# Windows (PowerShell)
ls $env:USERPROFILE\.claude\skills\prompt-architect\

# Should see:
# - SKILL.md (required)
# - references/ (framework docs)
# - assets/ (templates)
# - scripts/ (analysis tools)
  1. Restart Claude Code

The skill will be automatically loaded on restart.

Verification

Test the skill with:

"Help me improve this prompt: analyze some data"

You should see the skill activate with prompt analysis and framework recommendations.


Updating

Update via npm

If installed globally:

# Update to latest version
npm update -g @ckelsoe/claude-skill-prompt-architect

# Or reinstall
npm install -g @ckelsoe/claude-skill-prompt-architect@latest

# Restart Claude Code

If installed locally:

# Update package
npm update @ckelsoe/claude-skill-prompt-architect

# Reinstall skill
npx prompt-architect-install

# Restart Claude Code

Check current version:

npm list -g @ckelsoe/claude-skill-prompt-architect

Note: npm packages do not auto-update. You must manually run the update command to get the latest version.


Usage

Basic Usage

Simply ask Claude Code to improve any prompt:

"Help me improve this prompt: [your prompt here]"

Example Interaction

User: "Help me improve this prompt: create a function to process user data"

Skill: I'll analyze this prompt and help improve it.

**Analysis**:
- Clarity: 4/10 (basic goal stated)
- Specificity: 2/10 (what kind of data? what processing?)
- Context: 1/10 (no background)
- Completeness: 2/10 (missing many details)
- Structure: 3/10 (single sentence)
- **Overall: 2.4/10**

**Recommendation**: RISEN framework (multi-step process requiring methodology)

**Questions**:
1. What type of user data? (names, emails, profiles, etc.)
2. What processing is needed? (validation, transformation, storage?)
3. What's the tech stack? (Python, JavaScript, database type?)
4. Any constraints? (performance, security requirements?)
5. Expected error handling approach?

User: "It's a Python function processing CSV user registrations. Need to validate
emails, normalize names, and save to PostgreSQL. Should handle invalid data gracefully."

Skill: Perfect! Let me apply the RISEN framework...

[Improved prompt generated with clear role, instructions, steps, end goal, and constraints]

Advanced Usage

Request Specific Framework:

"Use the CO-STAR framework to improve: write marketing copy"

Iterate on Existing Prompts:

"Review this prompt and suggest improvements: [existing structured prompt]"

Switch Frameworks:

"Try using TIDD-EC instead for more explicit guidance"

Framework Selection Guide

Decision Tree

Are you transforming existing content (rewrite, refactor, convert)?
├─ YES → BAB (Before, After, Bridge)
└─ NO ↓

Is it an agentic/tool-use task?
├─ YES → ReAct (Thought-Action-Observation cycles)
└─ NO ↓

Is it a decision between multiple approaches?
├─ YES → Tree of Thought (branching exploration)
└─ NO ↓

Is it content/writing focused with audience and tone?
├─ YES → CO-STAR
└─ NO ↓

Is it a multi-step process with methodology?
├─ YES → RISEN
└─ NO ↓

Is it a data transformation (input → output)?
├─ YES → RISE-IE
└─ NO ↓

Does it need reference examples?
├─ YES → RISE-IX
└─ NO ↓

Does it need explicit dos/don'ts?
├─ YES → TIDD-EC
└─ NO ↓

Is it complex reasoning (one clear path)?
├─ YES → Chain of Thought
└─ NO ↓

Does it need iterative refinement?
├─ YES → Chain of Density
└─ NO ↓

Is it a numerical/calculation problem (zero-shot)?
├─ YES → Plan-and-Solve PS+
└─ NO ↓

Is it compositional / multi-hop (answer A needed before B)?
├─ YES → Least-to-Most
└─ NO ↓

Does it need first-principles reasoning?
├─ YES → Step-Back Prompting
└─ NO ↓

Is it structured long-form content (multiple semi-independent sections)?
├─ YES → Skeleton of Thought
└─ NO ↓

Do you need to recover a prompt from an existing output?
├─ YES → RPEF (Reverse Prompt Engineering)
└─ NO ↓

Do you need to clarify requirements before starting?
├─ YES → Reverse Role Prompting (AI interviews you first)
└─ NO ↓

Do you need to critique, stress-test, or verify something?
├─ General quality improvement → Self-Refine
├─ Align to explicit principle → CAI Critique-Revise
├─ Find opposing argument → Devil's Advocate
├─ Find failure modes → Pre-Mortem
└─ Verify conditions weren't missed → RCoT

Is it a simple task? Choose by primary driver:
├─ Expert role matters most → RTF
├─ Situational context matters most → CTF
├─ Need role + context + outcome bar → RACE
├─ Business deliverable with KPIs → BROKE
├─ Want multiple variants to compare → CRISPE
├─ Have explicit rules/constraints → CARE
└─ Ultra-minimal, one-off → APE

Quick Reference

| Your Task | Recommended Framework | Why | |-----------|----------------------|-----| | Write blog post | CO-STAR | Audience and tone critical | | Code review process | RISEN | Multi-step with constraints | | Analyze CSV data | RISE-IE | Clear input→output transformation | | Generate with examples | RISE-IX | Need reference samples | | Generate secure code | TIDD-EC | Explicit dos/don'ts needed | | Debug logic error | Chain of Thought | Requires reasoning steps | | Compress explanation | Chain of Density | Iterative refinement | | Simple conversion | RTF | Straightforward, expertise-driven | | Mid-project update | CTF | Background is the key driver | | Summarize a meeting | APE | Ultra-minimal, one-off | | Refactor existing code | BAB | Clear before/after transformation | | Rewrite copy for new audience | BAB | Current content → desired state | | Architecture decision | Tree of Thought | Multiple approaches to compare | | Choose database tech | Tree of Thought | Trade-offs need systematic analysis | | Agentic research task | ReAct | Tool use with iterative reasoning | | Expert review with context | RACE | Role + background + outcome bar | | Financial calculation | Plan-and-Solve (PS+) | Zero-shot + variable extraction | | Multi-hop technical question | Least-to-Most | Dependencies must be solved in order | | Architecture principle question | Step-Back | Abstract to first principles first | | Write structured docs/report | Skeleton of Thought | Outline first, expand second | | Marketing copy A/B options | CRISPE | Experiment component generates variants | | Business strategy with KPIs | BROKE | Key Results + Evolve self-critique | | Healthcare/compliance content | CARE | Explicit rules and quality standards | | Improve any output quality | Self-Refine | Multi-dimensional feedback + refine loop | | Enforce plain language standard | CAI Critique-Revise | Principle-based critique | | Stress-test a strategy | Devil's Advocate | Strongest opposing argument | | Project risk analysis | Pre-Mortem | Assumed failure → backward causes | | Multi-condition logic verification | RCoT | Backward reconstruction cross-check | | Lost/need to recover a prompt | RPEF | Output → reconstructed template | | Requirements unclear upfront | Reverse Role Prompting | AI interviews you first |


Project Structure

prompt-architect/
│
├── README.md                          # This file
├── LICENSE                            # MIT License
│
└── prompt-architect/                  # The skill
    ├── SKILL.md                       # Core skill instructions (5 KB)
    │
    ├── scripts/                       # Analysis Utilities
    │   ├── framework_analyzer.py      # Framework recommendation logic
    │   └── prompt_evaluator.py        # Quality scoring system
    │
    ├── references/                    # Framework Documentation
    │   └── frameworks/                # Loaded on-demand
    │       ├── co-star.md             # CO-STAR reference (600+ lines)
    │       ├── risen.md               # RISEN reference (600+ lines)
    │       ├── rise.md                # RISE (IE/IX) reference (700+ lines)
    │       ├── tidd-ec.md             # TIDD-EC reference (600+ lines)
    │       ├── ctf.md                 # CTF reference
    │       ├── rtf.md                 # RTF reference (500+ lines)
    │       ├── ape.md                 # APE reference
    │       ├── bab.md                 # BAB reference
    │       ├── race.md                # RACE reference
    │       ├── crispe.md              # CRISPE reference
    │       ├── broke.md               # BROKE reference
    │       ├── care.md                # CARE reference
    │       ├── tree-of-thought.md     # Tree of Thought reference
    │       ├── react.md               # ReAct reference
    │       ├── skeleton-of-thought.md # Skeleton of Thought reference (ICLR 2024)
    │       ├── step-back.md           # Step-Back Prompting reference (Google DeepMind)
    │       ├── least-to-most.md       # Least-to-Most reference (Google Brain)
    │       ├── plan-and-solve.md      # Plan-and-Solve PS+ reference (ACL 2023)
    │       ├── chain-of-thought.md    # CoT reference (500+ lines)
    │       ├── chain-of-density.md    # CoD reference (500+ lines)
    │       ├── self-refine.md         # Self-Refine (NeurIPS 2023)
    │       ├── cai-critique-revise.md # CAI Critique-Revise (Anthropic 2022)
    │       ├── devils-advocate.md     # Devil's Advocate (ACM IUI 2024)
    │       ├── pre-mortem.md          # Pre-Mortem (Gary Klein)
    │       ├── rcot.md                # Reverse Chain-of-Thought
    │       ├── rpef.md                # Reverse Prompt Engineering (EMNLP 2025)
    │       └── reverse-role.md        # AI-Led Interview / FATA (arXiv 2025)
    │
    └── assets/
        └── templates/                 # Framework Templates
            ├── co-star_template.txt
            ├── risen_template.txt
            ├── rise-ie_template.txt
            ├── rise-ix_template.txt
            ├── tidd-ec_template.txt
            ├── ctf_template.txt
            ├── rtf_template.txt
            ├── ape_template.txt
            ├── bab_template.txt
            ├── race_template.txt
            ├── crispe_template.txt
            ├── broke_template.txt
            ├── care_template.txt
            ├── tree-of-thought_template.txt
            ├── react_template.txt
            ├── skeleton-of-thought_template.txt
            ├── step-back_template.txt
            ├── least-to-most_template.txt
            ├── plan-and-solve_template.txt
            ├── chain-of-thought_template.txt
            ├── chain-of-density_template.txt
            ├── self-refine_template.txt
            ├── cai-critique-revise_template.txt
            ├── devils-advocate_template.txt
            ├── pre-mortem_template.txt
            ├── rcot_template.txt
            ├── rpef_template.txt
            ├── reverse-role_template.txt
            └── hybrid_template.txt

Core Components:

  • SKILL.md - Main skill logic and instructions (intent-based selection system)
  • 27 Framework Docs - Complete references with examples
  • 29 Templates - Ready-to-use structures (27 frameworks + hybrid + chain variants)
  • 2 Python Scripts - Analysis and scoring utilities

Documentation

Framework References

Detailed documentation for each framework in prompt-architect/references/frameworks/:

  • co-star.md - Context, Objective, Style, Tone, Audience, Response
  • risen.md - Role, Instructions, Steps, End goal, Narrowing
  • rise.md - Dual variants (IE & IX)
  • tidd-ec.md - Task, Instructions, Do, Don't, Examples, Context
  • ctf.md - Context, Task, Format
  • rtf.md - Role, Task, Format
  • ape.md - Action, Purpose, Expectation
  • bab.md - Before, After, Bridge
  • race.md - Role, Action, Context, Expectation
  • crispe.md - Capacity+Role, Insight, Instructions, Personality, Experiment
  • broke.md - Background, Role, Objective, Key Results, Evolve
  • care.md - Context, Ask, Rules, Examples
  • tree-of-thought.md - Branching multi-path exploration
  • react.md - Reasoning + Acting (agentic tool use)
  • skeleton-of-thought.md - Skeleton-first parallel expansion (ICLR 2024)
  • step-back.md - Abstract to principles first (Google DeepMind, ICLR 2024)
  • least-to-most.md - Sequential subproblem solving (Google Brain, ICLR 2023)
  • plan-and-solve.md - Zero-shot PS+ trigger (ACL 2023)
  • chain-of-thought.md - Step-by-step reasoning
  • chain-of-density.md - Iterative refinement
  • self-refine.md - Generate → Feedback → Refine (NeurIPS 2023)
  • cai-critique-revise.md - Principle-based critique + revision (Anthropic 2022)
  • devils-advocate.md - Strongest opposing argument (ACM IUI 2024)
  • pre-mortem.md - Assumed failure → backward cause analysis (Gary Klein)
  • rcot.md - Reverse Chain-of-Thought verification
  • rpef.md - Reverse Prompt Engineering (EMNLP 2025)
  • reverse-role.md - AI-Led Interview / FATA (arXiv 2025)

Contributing

Contributions are welcome! Here's how you can help:

Ways to Contribute

  1. Add New Frameworks - Implement additional prompting frameworks
  2. Improve Existing Frameworks - Enhance documentation and examples
  3. Bug Fixes - Report and fix issues
  4. Documentation - Improve guides and examples
  5. Testing - Test with different use cases and report findings

Adding New Frameworks

  1. Research and validate the framework has backing
  2. Create comprehensive framework reference (see existing in references/frameworks/)
  3. Add template file in assets/templates/
  4. Update SKILL.md framework selection logic
  5. Test with multiple use cases
  6. Submit pull request with examples and rationale

Contribution Guidelines

File Structure:

New framework contributions should include:
- references/frameworks/your-framework.md (600+ lines with examples)
- assets/templates/your-framework_template.txt
- Updates to SKILL.md framework selection
- Examples in documentation

Documentation Standards:

  • Clear component definitions
  • Multiple use case examples
  • Selection criteria explanation
  • Before/after comparisons
  • Research citations

Testing Requirements:

  • Test with 3-5 diverse prompts
  • Verify scoring improves
  • Check framework selection logic
  • Validate template structure

Pull Request Process

  1. Fork the repository
  2. Create a feature branch (git checkout -b feature/new-framework)
  3. Make your changes following the guidelines
  4. Test thoroughly with multiple examples
  5. Update relevant documentation
  6. Commit with clear messages
  7. Push to your fork
  8. Submit a Pull Request with:
    • Clear description of changes
    • Examples of improvements
    • Test results
    • Documentation updates

License

This project is licensed under the MIT License - see the LICENSE file for details.

License Summary

Permissions:

  • Commercial use
  • Modification
  • Distribution
  • Private use

Conditions:

  • License and copyright notice must be included

Limitations:

  • No liability
  • No warranty

Support

Getting Help

  • Issues - Report bugs or request features via GitHub Issues
  • Documentation - Check framework references in references/frameworks/

FAQ

Q: Why isn't the skill activating? A: Ensure the skill folder is in the correct location (~/.claude/skills/prompt-architect/) and restart Claude Code.

Q: Can I use this without Claude Code? A: Yes! Include SKILL.md content in your Claude API system prompt.

Q: Which framework should I use? A: The skill will recommend one, but check the Framework Selection Guide above.

Q: Can I add my own framework? A: Yes! See the Contributing section for guidelines.

Q: Does this work with other LLMs? A: The frameworks are universal, but the skill is optimized for Claude.


Transforming your prompts with research-backed frameworks