npm package discovery and stats viewer.

Discover Tips

  • General search

    [free text search, go nuts!]

  • Package details

    pkg:[package-name]

  • User packages

    @[username]

Sponsor

Optimize Toolset

I’ve always been into building performant and accessible sites, but lately I’ve been taking it extremely seriously. So much so that I’ve been building a tool to help me optimize and monitor the sites that I build to make sure that I’m making an attempt to offer the best experience to those who visit them. If you’re into performant, accessible and SEO friendly sites, you might like it too! You can check it out at Optimize Toolset.

About

Hi, 👋, I’m Ryan Hefner  and I built this site for me, and you! The goal of this site was to provide an easy way for me to check the stats on my npm packages, both for prioritizing issues and updates, and to give me a little kick in the pants to keep up on stuff.

As I was building it, I realized that I was actually using the tool to build the tool, and figured I might as well put this out there and hopefully others will find it to be a fast and useful way to search and browse npm packages as I have.

If you’re interested in other things I’m working on, follow me on Twitter or check out the open source projects I’ve been publishing on GitHub.

I am also working on a Twitter bot for this site to tweet the most popular, newest, random packages from npm. Please follow that account now and it will start sending out packages soon–ish.

Open Software & Tools

This site wouldn’t be possible without the immense generosity and tireless efforts from the people who make contributions to the world and share their work via open source initiatives. Thank you 🙏

© 2026 – Pkg Stats / Ryan Hefner

uer-mcp

v4.2.2

Published

[Hackathon Proof-of-Concept] Universal Expert Registry - Multi-provider LLM gateway, S3-compatible storage, and MCP tool orchestration. Tested with Anthropic, Cerebras, OpenAI, Gemini, LM Studio, Ollama.

Readme

Universal Expert Registry

npm version npm npm bundle size License: MIT

Multi-Provider LLM Gateway • S3-Compatible Storage • MCP Tool Orchestration

⚠️ Development Status: This is a hackathon proof-of-concept. While the architecture supports 100+ LLM providers via LiteLLM, only a subset (Anthropic, Cerebras, OpenAI, Gemini, LM Studio, Ollama) have been extensively tested. Version numbers track feature implementation progress, not production readiness. If you encounter issues with other providers, please open an issue.


Standard config works in most MCP clients:

Quick Start: Get a free Cerebras API key at cloud.cerebras.ai/platform under apikeys or use LM Studio (100% free, local)

{
  "mcpServers": {
    "uer": {
      "command": "npx",
      "args": [
        "-y",
        "uer-mcp@latest"
      ],
      "env": {
        "CEREBRAS_API_KEY": "<YOUR_TOKEN>",
        "GEMINI_API_KEY": "<YOUR_TOKEN>",
        "LM_STUDIO_API_BASE": "http://localhost:1234/v1"
      }
    }
  }
}

An actual developer setup for Claude could look like (docker exists and minio is used for that):

{
  "mcpServers": {
    "uer": {
      "command": "uv",
      "args": ["--directory", "C:\\Users\\margu\\UER", "run", "python", "-m", "uer.server"],
      "env": {
        "GEMINI_API_KEY": "AIzaSyAzXhhzgWzCBL7...",
        "ANTHROPIC_API_KEY": "sk-ant-api03--E1YU1bN0rdZjkJrBOiR...",
        "CEREBRAS_API_KEY":"csk-9we5kdvjc5efnnfefwhc6w...",
        "LM_STUDIO_API_BASE": "http://localhost:1234/v1"
      }
    }
  }
}

Storage Setup (Optional):

  • Quick start: Run docker-compose up -d to start MinIO locally
  • See Storage Configuration Options for detailed setup (Docker MinIO, AWS S3, or manual configuration)
  • Storage enables skills, templates, and behavior monitoring features

Storage is optional: This config works immediately for LLM and MCP features. Storage features activate automatically when MinIO is running. For storage/context features, see Storage Configuration Options below.

Required: Add at least one API key to the env section. See CONFIGURATION.md for all provider links and detailed setup.

For Claude Desktop, Goose, Codex, Amp, and other clients, see CONFIGURATION.md for detailed setup instructions.


An MCP server that provides:

  1. Multi-Provider LLM Access - Call 100+ LLM providers (Anthropic, OpenAI, Google, Azure, AWS Bedrock, local models) through LiteLLM
  2. MCP Tool Integration - Connect to other MCP servers for extended functionality
  3. S3-Compatible Storage - Store context and data in MinIO, AWS S3, or other S3-compatible backends
  4. Prompt Injection Detection - Basic content validation and security warnings

Why This Exists

MCP clients often need:

  • Multiple LLM providers - Different models for different tasks
  • Persistent storage - Save context between sessions
  • Tool integration - Connect to specialized MCP servers
  • Configuration flexibility - Support cloud and self-hosted solutions

UER provides:

  • Unified interface to 100+ LLM providers via LiteLLM
  • S3-compatible storage for context and data
  • MCP client for calling other MCP servers
  • Support for enterprise clouds (Azure, AWS, GCP) and self-hosted (Ollama, LM Studio)

Architecture

graph TB
    subgraph clients["MCP Clients"]
        A1["Cursor"]
        A2["Claude Desktop"]
        A3["ChatGPT"]
        A4["VS Code"]
        A5["JetBrains"]
    end

    subgraph uer["UER - Universal Expert Registry"]
        direction TB
        B["MCP Tools<br/>llm_call, mcp_call, put, get, delegate, search"]

        subgraph litellm["LiteLLM Gateway"]
            C1["100+ LLM providers"]
            C2["Model routing"]
            C3["Error handling"]
            C4["Response formatting"]
        end

        subgraph store["Context Store"]
            D1["Local: MinIO"]
            D2["Cloud: AWS S3, Azure, NetApp"]
        end

        B --> litellm
        B --> store
    end

    subgraph providers["LLM Providers"]
        E1["Anthropic"]
        E2["OpenAI"]
        E3["Google"]
        E4["Azure"]
        E5["AWS Bedrock"]
        E6["Local: Ollama"]
    end

    subgraph mcpservers["MCP Servers"]
        F1["Filesystem"]
        F2["PostgreSQL"]
        F3["Slack"]
        F4["Browser"]
        F5["GitHub"]
        F6["1000+ more..."]
    end

    subgraph knowledge["Knowledge Sources"]
        G1["Context7"]
        G2["Company docs"]
        G3["Guidelines"]
        G4["Standards"]
    end

    clients -->|MCP Protocol| B
    litellm --> providers
    litellm --> mcpservers
    litellm --> knowledge

Key Features

1. Universal LLM Access via LiteLLM

Call any LLM with a single interface:

# All use the same interface - just change the model string
llm_call(model="anthropic/claude-sonnet-4-5-20250929", messages=[...])
llm_call(model="openai/gpt-5.2", messages=[...])
llm_call(model="gemini/gemini-3-flash-preview", messages=[...])
llm_call(model="bedrock/anthropic.claude-3-sonnet", messages=[...])
llm_call(model="azure/gpt-4-deployment", messages=[...])
llm_call(model="ollama/llama3.1:8b-instruct-q4_K_M", messages=[...])

Features included:

  • Unified interface across providers
  • Support for cloud and self-hosted models
  • Automatic model detection and caching
  • Error handling and response formatting

2. MCP Tool Integration

Connect to any MCP server:

# List available MCP tools
search(type="mcp")

# Call MCP tools directly
mcp_call(server="filesystem", tool="read_file", args={"path": "/data/report.txt"})
mcp_call(server="postgres", tool="query", args={"sql": "SELECT * FROM users"})
mcp_call(server="context7", tool="search", args={"query": "LiteLLM API reference"})

3. S3-Compatible Storage

Store data in S3-compatible backends:

# Store data in MinIO, AWS S3, or other S3-compatible storage
storage_put(
    key="analysis/doc_001.json",
    content={"content": large_document},
    bucket="uer-context"
)

# Retrieve data
data = storage_get(
    key="analysis/doc_001.json",
    bucket="uer-context"
)

Storage backends:

  • Local: MinIO (S3-compatible, Docker-based)
  • Cloud: AWS S3, Azure Blob Storage, NetApp StorageGRID
  • Features: Versioning, WORM compliance, Jinja2 templates, Claude Skills API support

See docs/ADR-002-S3-Storage-Architecture.md for details.

Storage Configuration Options

UER supports three deployment scenarios for storage:

Option 1: Docker MinIO (Recommended for Development)

If you have Docker installed, start MinIO with one command:

docker-compose up -d

This starts MinIO on localhost:9000 with default credentials (minioadmin/minioadmin). UER will automatically connect and create the required buckets (uer-context, uer-skills, uer-templates) on first use.

Access the MinIO console at http://localhost:9001 to browse stored objects.

Option 2: Custom S3-Compatible Storage

For production or if you don't use Docker, configure your own S3-compatible storage:

{
  "mcpServers": {
    "uer": {
      "command": "npx",
      "args": ["uer-mcp@latest"],
      "env": {
        "GEMINI_API_KEY": "your-key-here",
        "STORAGE_BACKEND": "minio",
        "MINIO_ENDPOINT": "your-minio-server.com:9000",
        "MINIO_ACCESS_KEY": "your-access-key",
        "MINIO_SECRET_KEY": "your-secret-key",
        "MINIO_SECURE": "true"
      }
    }
  }
}

Supports any S3-compatible storage:

  • MinIO (self-hosted)
  • AWS S3 (use S3_ENDPOINT, S3_ACCESS_KEY, S3_SECRET_KEY, S3_REGION)
  • NetApp StorageGRID
  • Wasabi, Backblaze B2, DigitalOcean Spaces

Option 3: Disabled Storage (LLM/MCP Only)

If you only need LLM and MCP features without storage:

{
  "mcpServers": {
    "uer": {
      "command": "npx",
      "args": ["uer-mcp@latest"],
      "env": {
        "GEMINI_API_KEY": "your-key-here",
        "STORAGE_ENABLED": "false"
      }
    }
  }
}

With storage disabled:

  • llm_call - Call any LLM
  • mcp_call, mcp_list_tools, mcp_servers - MCP orchestration
  • ❌ Storage tools (storage_put, storage_get, etc.) - Not available
  • ❌ Skills tools (skill_create, skill_get, etc.) - Not available
  • ❌ Template tools (template_render, etc.) - Not available

The server will start successfully without storage, and LLMs won't see storage-related tools in their tool list.

4. Prompt Injection Detection

Basic content validation and security warnings:

# Detects potential prompt injection patterns
# Provides risk assessment and warnings
# Helps identify suspicious content in user inputs

Usage

Test Your Setup

Try this in Claude Desktop:

"Use the llm_call tool to call Gemini 3 Flash and ask it to explain what an MCP server is in one sentence."

Expected behavior:

  • Claude will use the llm_call tool
  • Call gemini/gemini-3-flash-preview
  • Return Gemini's response

Example Usage Scenarios

1. Call Different LLMs:

User: "Use llm_call to ask Gemini what the capital of France is"
→ Calls gemini/gemini-3-flash-preview
→ Returns: "Paris"

User: "Now ask Claude Sonnet the same question"
→ Calls anthropic/claude-sonnet-4-5-20250929
→ Returns: "Paris"

2. Compare LLM Responses:

User: "Ask both Gemini and Claude Sonnet to write a haiku about programming"
→ Uses llm_call twice with different models
→ Returns both haikus for comparison

3. Store and Retrieve Data:

User: "Store this configuration in S3"
→ storage_put(key="config/settings.json", content={...})
→ Returns: Confirmation with storage details

User: "Retrieve the configuration"
→ storage_get(key="config/settings.json")
→ Returns: Configuration data

Troubleshooting

"MCP server not found" or "No tools available"

  1. Check that claude_desktop_config.json is in the correct location
  2. Verify the --directory path is correct (use absolute path)
  3. Ensure you've restarted Claude Desktop after configuration
  4. Check Claude Desktop logs: %APPDATA%\Claude\logs\ (Windows) or ~/Library/Logs/Claude/ (Mac)

"API key invalid" errors

  1. Verify your API key is correct and active
  2. Check you're using the right key for the right provider
  3. For Gemini, ensure the key starts with AIza
  4. For Anthropic, ensure the key starts with sk-ant-
  5. For OpenAI, ensure the key starts with sk-

"Model not found" errors

  1. Ensure you have an API key configured for that provider
  2. Check the model name is correct (use LiteLLM format: provider/model)
  3. Verify the model is available in your region/tier

Tools Reference

| Tool | Description | |------|-------------| | llm_call | Call any LLM via LiteLLM (100+ providers) | | llm_list_models | List available models from configured providers | | llm_config_guide | Get configuration help for LLM providers | | mcp_call | Call any configured MCP server tool | | mcp_list_tools | List available MCP tools | | mcp_servers | List configured MCP servers | | storage_put | Store data in S3-compatible storage | | storage_get | Retrieve data from storage | | storage_list | List stored objects | | storage_delete | Delete stored objects |

LiteLLM Integration

This project uses LiteLLM as the unified LLM gateway, providing:

  • 100+ LLM providers through single interface
  • Unified API format across all providers
  • Support for cloud and self-hosted models
  • Automatic model detection and caching
  • Error handling and response formatting

Provider & Model Discovery

Find supported providers and models:

  • 📖 PROVIDERS.md - Complete guide to LiteLLM provider integrations and configuration
  • 🌐 LiteLLM Provider Docs - Official documentation for all 100+ providers
  • 🔧 llm_list_models tool - Query available models from your configured providers
  • 🔧 llm_config_guide tool - Get configuration help for specific providers

Supported Providers (Examples)

| Provider | Model Examples | Testing Status | |----------|---------------|----------------| | Anthropic | anthropic/claude-sonnet-4-5-20250929, anthropic/claude-opus-4-5-20251101 | ✅ Tested | | Cerebras | cerebras/llama-3.3-70b, cerebras/qwen-3-235b-a22b-instruct-2507 | ✅ Tested | | OpenAI | openai/gpt-4o, openai/o3-mini | ✅ Tested | | Google | gemini/gemini-2.5-flash, gemini/gemini-2.0-flash-exp | ✅ Tested | | LM Studio | lm_studio/meta-llama-3.1-8b-instruct (local) | ✅ Tested | | Ollama | ollama/llama3.1:8b-instruct-q4_K_M (local) | ✅ Tested | | Azure | azure/gpt-4-deployment | ⚠️ Untested | | AWS Bedrock | bedrock/anthropic.claude-3-sonnet | ⚠️ Untested | | Cohere | cohere_chat/command-r-plus | ⚠️ Untested | | Together AI | together_ai/meta-llama/Llama-3-70b-chat-hf | ⚠️ Untested |

Testing Status:

  • Tested: Verified during development with live API queries and model caching
  • ⚠️ Untested: Supported via LiteLLM but not extensively tested. May require minor adjustments. Please report issues if you encounter problems.

Note: Model names change frequently. Use the discovery tools above to find current models.

Advanced Configuration

Multi-Instance Providers: LiteLLM supports multiple instances of the same provider (e.g., multiple Azure deployments). Configure via environment variables:

# Multiple Azure deployments
AZURE_API_KEY="key1"
AZURE_API_BASE="https://endpoint1.openai.azure.com"
AZURE_API_VERSION="2023-05-15"

# Use model format: azure/<deployment-name>
# Example: azure/gpt-4-deployment

Generic Provider Support: Any provider with a configured API key will be detected automatically. If we don't have a specific query implementation, example models will be provided. Supported providers include:

  • Cohere (COHERE_API_KEY)
  • Together AI (TOGETHERAI_API_KEY)
  • Replicate (REPLICATE_API_KEY)
  • Hugging Face (HUGGINGFACE_API_KEY)
  • And 90+ more - see LiteLLM docs

Fallback Chains: LiteLLM supports automatic fallbacks. Configure via model list:

# In your LLM call, specify fallback models
model="gpt-4o"  # Primary
fallbacks=["claude-sonnet-4-5", "gemini-2.5-flash"]  # Fallbacks

See PROVIDERS.md for detailed configuration examples.

Project Structure

UER/
├── README.md               # This file
├── ADR.plan.md            # Architecture Decision Record
├── TODO.md                # Implementation checklist
├── pyproject.toml
│
├── src/
│   ├── server.py          # MCP server entry point
│   ├── llm/
│   │   └── gateway.py     # LiteLLM wrapper
│   ├── mcp/
│   │   └── client.py      # MCP client for calling other servers
│   ├── storage/
│   │   ├── base.py        # S3-compatible storage protocol
│   │   ├── minio_backend.py  # MinIO backend (local)
│   │   ├── s3_backend.py     # AWS S3 backend (cloud)
│   │   ├── manager.py        # Storage manager
│   │   ├── skills.py         # Claude Skills API support
│   │   └── templates.py      # Jinja2 template rendering
│   ├── tools/
│   │   ├── llm_call.py    # LLM invocation tool
│   │   ├── mcp_call.py    # MCP tool invocation
│   │   ├── storage_tools.py  # put/get/list/delete
│   │   └── delegate.py    # Subagent delegation
│   └── models/
│       ├── storage.py     # Storage schemas (ObjectMetadata, Retention)
│       └── message.py     # Chat message schemas
│
└── config/
    └── litellm_config.yaml

Dependencies

[project]
dependencies = [
    "mcp>=1.0.0",
    "litellm>=1.77.0",
    "pydantic>=2.0.0",
    "httpx>=0.25.0",
]

Datasets & Testing

UER includes scripts to download and test manipulation detection datasets.

Quick Start: Download All Datasets

One command downloads everything:

python seed_datasets.py

This downloads:

  • WMDP Benchmark: 3,668 questions (Bio: 1,273, Chem: 408, Cyber: 1,987)
  • WildChat Sample: 10,000 real conversations (162 MB)
  • lm-evaluation-harness: Evaluation framework

Time: ~5-10 minutes depending on internet speed.

Run Tests

Test for Sandbagging:

cd context/scripts
python test_wmdp.py --model gemini/gemini-3-flash-preview --limit 50

Test for Sycophancy:

python test_sycophancy.py --models gemini

Results saved to: context/datasets/results/

Dataset Details

| Dataset | Size | Purpose | Location | |---------|------|---------|----------| | WMDP Benchmark | 3,668 questions (2.2 MB) | Sandbagging detection | context/datasets/wmdp_questions/ | | WildChat | 10k conversations (162 MB) | Real-world sycophancy | context/datasets/wildchat/ | | lm-evaluation-harness | Framework | Standard LLM evaluation | context/datasets/lm-evaluation-harness/ |

All datasets are gitignored. Run seed_datasets.py to download locally.

Hackathon Context

This project was built for the AI Manipulation Hackathon organized by Apart Research.

Event Details

  • Dates: January 9-11, 2026
  • Theme: Measuring, detecting, and defending against AI manipulation
  • Participants: 500+ builders worldwide
  • Prizes: $2,000 in cash prizes
  • Workshop: Winners present at IASEAI workshop in Paris (February 26, 2026)

The Challenge

AI systems are mastering deception, sycophancy, sandbagging, and psychological exploitation at scale, while our ability to detect, measure, and counter these behaviors remains dangerously underdeveloped. This hackathon brings together builders to prototype practical systems that address this critical AI safety challenge.

How UER Addresses AI Manipulation

The Universal Expert Registry provides infrastructure for:

  1. Multi-Model Testing - Compare responses across providers to detect inconsistencies and manipulation patterns
  2. Persistent Context - Track conversation history across sessions to identify behavioral shifts
  3. Tool Integration - Connect manipulation detection tools via MCP protocol
  4. Subagent Orchestration - Deploy specialized agents for red-teaming and safety testing
  5. Transparent Logging - Full visibility into LLM calls, costs, and behaviors

Team

The Risk Takers - Building practical tools for AI safety and transparency.

Hackathon Resources

The hackathon provides extensive research and tools for understanding AI manipulation. Below are key resources organized by track.

General Introduction

AI deception: A survey of examples, risks, and potential solutions (Park et al., 2024) Comprehensive survey defining AI deception and documenting manipulation via sycophancy, strategic behavior, and safety test evasion. Covers both near-term risks (fraud, election tampering) and long-term risks (losing control of AI systems).

Artificial Intelligence Index Report 2025 (Stanford HAI) Annual assessment of global AI progress including responsible AI and system risks. Provides regulatory trends and technical challenges context.

Towards Understanding Sycophancy in Language Models (Sharma et al., 2024) Documents that state-of-the-art AI assistants consistently exhibit sycophancy, with human preference data driving this behavior. Demonstrates inverse scaling: larger models become MORE sycophantic.

Track 1: Measurement & Evaluation

AI Sandbagging: Language Models can Strategically Underperform on Evaluations (van der Weij et al., 2024) Demonstrates GPT-4 and Claude 3 Opus strategically underperforming on dangerous capability evaluations. Models can "password-lock" capabilities and target specific scores.

Noise Injection Reveals Hidden Capabilities of Sandbagging Language Models (Tice et al., 2024) Novel sandbagging detection via weight noise injection. Provides practical, model-agnostic detection tool.

The WMDP Benchmark: Measuring and Reducing Malicious Use With Unlearning (Li et al., 2024) Dataset of 3,668 multiple-choice questions for evaluating hazardous knowledge. Widely used in sandbagging and capability evaluation research.

UER Application: Use llm_call to test multiple models against WMDP benchmark, comparing responses to detect sandbagging patterns.

Track 2: Real-World Analysis

Recent Frontier Models Are Reward Hacking (METR, 2025) Documents reward hacking in production-like settings: o3 hacked evaluation software, Claude models subverted tasks. Shows models demonstrate awareness their behavior violates intentions.

Reward Hacking in Reinforcement Learning (Lilian Weng, 2024) Comprehensive overview with real-world examples: summarization gaming ROUGE, coding tampering tests, recommendation algorithms optimizing engagement over welfare.

Reward hacking behavior can generalize across tasks (Denison et al., 2024) Training on reward hacking leads to 1.3x-2.0x increase on held-out datasets. Documents how manipulation generalizes.

UER Application: Use persistent context storage with put/get to track behavioral shifts across sessions and detect manipulation patterns.

Track 3: Mitigations

Chain of Thought Monitoring for Misbehavior (OpenAI, 2025) Chain-of-thought monitoring can detect reward hacking. GPT-4o can monitor stronger o3-mini model. Demonstrates transparency-based mitigation promise and pitfalls.

From shortcuts to sabotage: natural emergent misalignment from reward hacking (Anthropic, 2025) Reward hacking training causes downstream misalignment (alignment faking, sabotage). Shows interventions can break semantic links.

The WMDP Benchmark: Measuring and Reducing Malicious Use With Unlearning (Li et al., 2024) Develops RMU unlearning method removing hazardous knowledge while preserving general capabilities.

UER Application: Integrate mitigation tools via mcp_call to test interventions across multiple models simultaneously.

Track 4: Open Track (Multi-Agent & Emergent Behavior)

AgentVerse: Facilitating Multi-Agent Collaboration and Exploring Emergent Behaviors (Chen et al., 2024) Demonstrates emergent social behaviors in multi-agent systems: volunteer behaviors, conformity, destructive behaviors.

Emergence in Multi-Agent Systems: A Safety Perspective (2024) Investigates how specification insufficiency leads to emergent manipulative behavior when agents' learned priors conflict.

School of Reward Hacks: Hacking Harmless Tasks Generalizes to Misalignment (2024) Training on "harmless" reward hacking causes generalization to concerning behaviors including shutdown avoidance and alignment faking.

UER Application: Use delegate to orchestrate multi-agent studies with different models, tracking emergent manipulation behaviors via shared context.

Open Datasets & Tools

| Resource | Type | Link | |----------|------|------| | WMDP Benchmark | Dataset + Code | github.com/centerforaisafety/wmdp | | WildChat Dataset | 1M ChatGPT conversations | huggingface.co/datasets/allenai/WildChat | | lm-evaluation-harness | Evaluation framework | github.com/EleutherAI/lm-evaluation-harness | | METR Task Environments | Autonomous AI tasks | github.com/METR/task-standard | | TransformerLens | Interpretability library | github.com/neelnanda-io/TransformerLens | | AgentVerse Framework | Multi-agent collaboration | github.com/OpenBMB/AgentVerse | | Multi-Agent Particle Envs | OpenAI environments | github.com/openai/multiagent-particle-envs | | School of Reward Hacks | Training dataset | github.com/aypan17/reward-hacking | | NetLogo | Agent-based modeling | ccl.northwestern.edu/netlogo |

Project Scoping Advice

Based on successful hackathon retrospectives:

Focus on MVP, Not Production (2-day timeline):

  • Day 1: Set up environment, implement core functionality, basic pipeline
  • Day 2: Add 1-2 key features, create demo, prepare presentation

Use Mock/Simulated Data instead of real APIs:

  • Synthetic datasets (WMDP, WildChat, School of Reward Hacks)
  • Pre-recorded samples
  • Simulation environments (METR, AgentVerse)

Leverage Pre-trained Models - Don't train from scratch:

  • OpenAI/Anthropic APIs via UER's llm_call
  • Hugging Face pre-trained models
  • Existing detection tools as starting points

Clear Success Criteria - Define "working":

  • Benchmarks: Evaluates 3+ models on 50+ test cases with documented methodology
  • Detection: Identifies manipulation in 10+ examples with >70% accuracy
  • Analysis: Documents patterns across 100+ deployment examples with clear taxonomy
  • Mitigation: Demonstrates measurable improvement on 3+ manipulation metrics

Related Projects

License

MIT


Built for the AI Manipulation Hackathon by The Risk Takers team