@scrapeops/mcp
v1.0.1
Published
MCP server for ScrapeOps Proxy API integration. Features web browsing, anti-bot bypass, JavaScript rendering, screenshots, and LLM-powered structured data extraction for AI agents.
Readme
ScrapeOps MCP Server
A Model Context Protocol (MCP) server that exposes the full capabilities of the ScrapeOps Proxy API to LLM clients (Cursor, Claude Desktop, VS Code). Enable AI agents to browse the web, bypass anti-bots, render JavaScript, take screenshots, and perform structured data extraction autonomously.
Features
- 🌐 Web Browsing - Browse any webpage with proxy support
- 🌍 Geo-Targeting - Access websites from 12+ countries
- 🏠 Residential/Mobile Proxies - Higher success rates on challenging sites
- 🤖 Anti-Bot Bypass - Multiple bypass levels (Cloudflare, DataDome, PerimeterX, etc.)
- 🖼️ Screenshots - Capture visual snapshots of pages
- ⚡ JavaScript Rendering - Full headless browser capabilities
- 📊 Structured Extraction - LLM-powered data extraction with schemas
- 🔄 Auto-Extract - Domain-specific parsers for popular sites
Installation
Running with npx
env SCRAPEOPS_API_KEY=YOUR_API_KEY npx -y @scrapeops/mcpManual Installation
npm install -g @scrapeops/mcpConfiguration
Environment Variables
| Variable | Required | Description |
|----------|----------|-------------|
| SCRAPEOPS_API_KEY | Yes | Your ScrapeOps API key from scrapeops.io |
How It Works
The MCP server uses a simple, single-request approach:
- Basic Request: If no options are specified, a basic request is made (URL only)
- User-Specified Options: If options like
render_js,residential, orpremiumare specified, they are used directly - On Failure: Returns helpful error with suggestions - the AI/user decides what to do next
- No Auto-Retry: The server does not automatically retry with different options - this gives you full control
Example Flow:
- You ask: "Scrape https://example.com"
- Server makes basic request (no extra params)
- If it fails (403), returns error with suggestions: "Try with
residential: true" - You decide: "OK, scrape it with residential proxy"
- Server makes request with
residential: true
Running on Cursor
- Open Cursor Settings
- Go to Features > MCP Servers
- Click "+ Add new global MCP server"
- Enter the following configuration:
{
"mcpServers": {
"@scrapeops/mcp": {
"command": "npx",
"args": ["-y", "@scrapeops/mcp"],
"env": {
"SCRAPEOPS_API_KEY": "YOUR-API-KEY"
}
}
}
}Running on Claude Desktop
Add this to your claude_desktop_config.json:
{
"mcpServers": {
"@scrapeops/mcp": {
"command": "npx",
"args": ["-y", "@scrapeops/mcp"],
"env": {
"SCRAPEOPS_API_KEY": "YOUR_API_KEY_HERE"
}
}
}
}Running on VS Code
Add to your User Settings (JSON) via Ctrl + Shift + P → Preferences: Open User Settings (JSON):
{
"mcp": {
"inputs": [
{
"type": "promptString",
"id": "apiKey",
"description": "ScrapeOps API Key",
"password": true
}
],
"servers": {
"scrapeops": {
"command": "npx",
"args": ["-y", "@scrapeops/mcp"],
"env": {
"SCRAPEOPS_API_KEY": "${input:apiKey}"
}
}
}
}
}Running on Windsurf
Add to your ./codeium/windsurf/model_config.json:
{
"mcpServers": {
"@scrapeops/mcp": {
"command": "npx",
"args": ["-y", "@scrapeops/mcp"],
"env": {
"SCRAPEOPS_API_KEY": "YOUR_API_KEY"
}
}
}
}Running Local Server (HTTP/SSE Transport)
You can run the server locally as an HTTP/SSE server instead of using stdio transport. This is useful for development or custom deployments.
1. Start the server:
# Set the port for HTTP/SSE mode (required for local server)
export PORT=8080
export SCRAPEOPS_API_KEY=your-api-key-here
# Run the server
npm start
# or if you have the package installed globally
scrapeops-mcpThe server will start on http://localhost:8080/sse (or the port specified by the PORT environment variable).
Note: If PORT is not set, the server will run in stdio mode (for use with npx in MCP clients like Cursor). Set PORT explicitly to run as an HTTP/SSE server.
2. Configure Cursor to connect to the local server:
Edit your Cursor MCP configuration file (typically at ~/.cursor/mcp.json or in Cursor Settings):
{
"mcpServers": {
"@scrapeops/mcp": {
"url": "http://localhost:8080/sse",
"headers": {
"scrapeops-api-key": "your-api-key-here"
}
}
}
}Note: When using HTTP/SSE transport, you can pass the API key either:
- Via the
scrapeops-api-keyheader in the configuration (as shown above), or - Via the
SCRAPEOPS_API_KEYenvironment variable when starting the server
Available Tools
Tool 1: maps_web
General-purpose web browsing tool for reading pages, taking screenshots, and bypassing anti-bot protections.
Usage Examples:
// Simple page browse
{
"name": "maps_web",
"arguments": {
"url": "https://example.com"
}
}
// Screenshot from Germany with residential proxy
{
"name": "maps_web",
"arguments": {
"url": "https://example.de",
"country": "de",
"residential": true,
"screenshot": true
}
}
// Bypass Cloudflare protection
{
"name": "maps_web",
"arguments": {
"url": "https://protected-site.com",
"bypass_level": "cloudflare_level_2",
"residential": true,
"render_js": true
}
}Tool 2: extract_data
Structured data extraction using auto-parsing or LLM-powered extraction.
Usage Examples:
// Auto-extract from known domain
{
"name": "extract_data",
"arguments": {
"url": "https://www.amazon.com/dp/B09V3KXJPB",
"mode": "auto"
}
}
// LLM extraction for product page
{
"name": "extract_data",
"arguments": {
"url": "https://shop.example.com/product/123",
"mode": "llm",
"data_schema": "product_page",
"response_format": "json"
}
}
// Extract job listings with anti-bot bypass
{
"name": "extract_data",
"arguments": {
"url": "https://careers.example.com/jobs",
"mode": "llm",
"data_schema": "job_search_page",
"bypass_level": "generic_level_2",
"render_js": true
}
}User Stories
The Visual Debugger
"User complains a site looks broken in Germany. The AI calls
maps_web(url='...', country='de', screenshot=true). The user sees the actual screenshot of the site rendered via a German residential IP."
The Efficient Scraper
"User needs pricing data. Instead of fetching HTML and parsing it (wasting tokens), the AI calls
extract_data(url='...', mode='llm', data_schema='product_page'). ScrapeOps handles the heavy lifting, and the AI just displays the final JSON."
The Bypass Expert
"The AI tries to access a site and gets blocked. It automatically retries the request using
maps_webwithbypass_level='generic_level_3'andresidential=trueto overcome the blockage."
System Configuration
The server includes configurable retry parameters with exponential backoff:
const RETRY_CONFIG = {
maxAttempts: 1,
initialDelay: 1000,
};Retry Behavior:
- Network errors are retried once regardless of maxAttempts setting
- To enable retries, set
SCRAPEOPS_RETRY_MAX_ATTEMPTSenvironment variable
Custom Configuration Example:
# Enable retries with 3 attempts
export SCRAPEOPS_RETRY_MAX_ATTEMPTS=3
export SCRAPEOPS_RETRY_INITIAL_DELAY=1000Error Handling
| Status Code | Error | Resolution |
|-------------|-------|------------|
| 401 | Invalid API Key | Check your SCRAPEOPS_API_KEY environment variable |
| 403 | Forbidden | Target website blocking request - consider using advanced parameters |
| 404 | Not Found | Verify the URL is correct |
| 429 | Rate Limited | Too many requests - wait before retrying (NOT auto-retried) |
| 500 | Server Error | Automatically retried up to 3 times with exponential backoff |
| 502/503 | Gateway/Service Error | Temporary issue - NOT auto-retried |
Development
# Install dependencies
npm install
# Build
npm run build
# Run locally (stdio transport)
SCRAPEOPS_API_KEY=your-key npm start
# Run tests
npm testAPI Reference
Base URL: https://proxy.scrapeops.io/v1/
Authentication: Query parameter ?api_key=... (managed via server-side environment variables)
For full API documentation, visit ScrapeOps Documentation.
License
MIT License - see LICENSE file for details.
