json-mcp-filter
v1.1.0
Published
MCP server providing JSON schema generation and filtering tools using quicktype
Maintainers
Readme
JSON MCP Filter
A Model Context Protocol (MCP) server that provides JSON schema generation and filtering tools for both local files and remote HTTP/HTTPS endpoints. This server uses quicktype to convert JSON samples into TypeScript type definitions and offers JSON data filtering capabilities.
Particularly helpful for JSON files that are on the larger side which contains data you don't want included in your LLM context. Perfect for filtering API responses and extracting only relevant data.
Features
- JSON Schema Generation: Convert JSON files into TypeScript type definitions using quicktype-core
- JSON Filtering: Extract specific fields from JSON data using shape-based filtering
- Remote File Support: Fetch and process JSON from HTTP/HTTPS URLs and APIs
- Content Size Protection: Automatic handling of large responses (up to 50MB)
- MCP Integration: Seamlessly integrates with Claude Desktop and Claude Code
- Type Safety: Built with TypeScript and includes comprehensive error handling
- Enhanced Error Messages: Clear, actionable error messages for common issues
Tools Provided
json_schema
Generates TypeScript type definitions from JSON files or remote URLs.
Parameters:
filePath: Local file path or HTTP/HTTPS URL to the JSON data
Examples:
Local file:
json_schema({ filePath: "./data.json" })Remote JSON file:
json_schema({ filePath: "https://api.example.com/config.json" })API endpoint:
json_schema({ filePath: "https://jsonplaceholder.typicode.com/users" })Sample Input:
{"name": "John", "age": 30, "city": "New York"}Sample Output:
export interface GeneratedType {
name: string;
age: number;
city: string;
}json_filter
Extracts specific fields from JSON data using a shape definition. Perfect for filtering large API responses to include only relevant data.
Features:
- Automatic Chunking: When filtered data exceeds 400KB, the tool automatically splits results into manageable chunks
- Line-Based Pagination: Uses simple line-based splitting
- Default: Defaults to chunk 0 when no chunk index is specified
Parameters:
filePath: Local file path or HTTP/HTTPS URL to the JSON datashape: Shape object defining which fields to extractchunkIndex(optional): Index of chunk to retrieve (0-based). Only needed for large datasets that exceed 400KB after filtering
Chunking Behavior:
- Small data (≤400KB): Returns all filtered data in one response
- Large data (>400KB): Automatically chunks data and returns the requested chunk with metadata
- No configuration options (for now): 400KB threshold is fixed
json_dry_run
Analyzes JSON data size and provides chunking recommendations. Returns size information in bytes for each specified field and calculates how many chunks would be needed for json_filter.
Parameters:
filePath: Local file path or HTTP/HTTPS URL to the JSON datashape: Shape object defining what to analyze for size
Features:
- Filtered Size Calculation: Shows exact size after applying the shape filter
- Chunking Recommendations: Indicates how many 400KB chunks would be needed
Use Cases:
- Determine if chunking will be needed for a specific shape
- Compare relative sizes of different JSON sections
Examples:
Analyze data size:
json_dry_run({
filePath: "https://api.example.com/users",
shape: {"name": true, "email": true, "posts": {"title": true}}
})Sample Output:
Total file size: 2.4 MB (2,517,847 bytes)
Filtered size: 245.8 KB (251,847 bytes)
Recommended chunks: 1
Size breakdown:
{
"name": 1205,
"email": 1834,
"posts": {
"title": 5692
}
}json_filter Examples
Filter API response (small data):
json_filter({
filePath: "https://jsonplaceholder.typicode.com/users",
shape: {"name": true, "email": true, "address": {"city": true}}
})
# Returns: All filtered data in one response (no chunking needed)Filter large dataset (automatic chunking):
# Step 1: Check size first
json_dry_run({
filePath: "./large-dataset.json",
shape: {"users": {"id": true, "name": true}}
})
# Returns: "Filtered size: 2.1 MB, Recommended chunks: 6"
# Step 2: Get first chunk (default)
json_filter({
filePath: "./large-dataset.json",
shape: {"users": {"id": true, "name": true}}
})
# Returns: Chunk 0 data + metadata {"chunkIndex": 0, "totalChunks": 6}
# Step 3: Get subsequent chunks
json_filter({
filePath: "./large-dataset.json",
shape: {"users": {"id": true, "name": true}},
chunkIndex: 1
})
# Returns: Chunk 1 data + metadata {"chunkIndex": 1, "totalChunks": 6}Shape Examples:
// Extract single field
{"name": true}
// Extract multiple fields
{"name": true, "age": true}
// Extract nested fields
{"user": {"name": true, "email": true}}
// Extract from arrays (applies to each item)
{"users": {"name": true, "age": true}}
// Complex nested extraction
{
"results": {
"name": {"first": true, "last": true},
"email": true,
"location": {"city": true, "country": true}
}
}⚠️ Security Disclaimer
IMPORTANT: Remote Data Fetching
This tool can fetch data from remote HTTP/HTTPS URLs. Users are responsible for:
- Verifying URLs before submission - Ensure URLs point to legitimate, safe endpoints
- Data validation - Review data sources and content before processing
- Rate limiting - Respect API rate limits and terms of service of external services
- Content safety - This tool does not validate the safety or appropriateness of remote content
The repository maintainers are not responsible for:
- Data fetched from external URLs
- Privacy implications of remote requests
- Malicious or inappropriate content from third-party sources
- API abuse or violations of third-party terms of service
Recommendations:
- Only use trusted, public APIs and data sources
- Verify URLs are legitimate before processing
- Be cautious with internal/localhost URLs in shared environments
- Review API documentation and terms of service before use
Installation
Quick Start (Recommended)
# Using npx (no installation required)
npx json-mcp-filter@latest
# Or install globally
npm install -g json-mcp-filter@latest
json-mcp-serverFrom Source
- Clone this repository:
git clone <repository-url>
cd json-mcp-filter- Install dependencies:
npm install- Build the server:
npm run buildSetup for Claude Desktop
Add this server to your Claude Desktop configuration file:
macOS
{
"mcpServers": {
"json-mcp-filter": {
"command": "node",
"args": ["/path/to/json-mcp-filter/build/index.js"]
}
}
}Setup for Claude Code
Add this server to your Claude Code MCP settings:
Add a new server with:
- Name:
json-mcp-filter - Command:
node - Args:
["/path/to/json-mcp-filter/build/index.js"]
Or, use the npx method for easier setup:
{
"mcpServers": {
"json-mcp-filter": {
"command": "npx",
"args": ["-y", "json-mcp-filter@latest"]
}
}
}Or claude mcp add json-mcp-filter node /path/to/json-mcp-filter/build/index.js
Development
Scripts
npm run build- Compile TypeScript and make executablenpm run start- Run the compiled servernpm run inspect- Run with MCP inspector for debuggingnpx tsc --noEmit- Type check without emitting files
Testing
Test the server using the MCP inspector:
npm run inspectThis will start the server with the MCP inspector interface for interactive testing.
Project Structure
src/
index.ts # Main server implementation with tools
strategies/ # Strategy pattern for data ingestion
JsonIngestionStrategy.ts # Abstract strategy interface
LocalFileStrategy.ts # Local file system access
HttpJsonStrategy.ts # HTTP/HTTPS URL fetching
context/
JsonIngestionContext.ts # Strategy management and selection
types/
JsonIngestion.ts # Shared type definitions
test/
test.json # Sample JSON files for testing
build/ # Compiled TypeScript outputError Handling
The server includes comprehensive error handling for:
Local File Errors
- File not found errors
- Invalid JSON format
- File permission issues
Remote URL Errors
- Network errors - Connection failures, timeouts
- Authentication errors - 401/403 responses with guidance
- Server errors - 500-series responses with retry suggestions
- Content size errors - Automatic rejection of responses over 50MB
- Format errors - Non-JSON content with format-specific guidance
- Rate limiting - 429 responses with wait instructions
Processing Errors
- Quicktype processing errors
- Shape filtering errors
- Invalid URL format errors
All errors include actionable messages and debugging information to help resolve issues quickly.
Performance & File Size Handling
Expected Performance
The JSON MCP tools are optimized for real-world JSON processing with the following performance characteristics:
| File Size | Processing Time | |-----------|-----------------| | < 100 KB | < 10ms | | 100 KB - 1 MB | 10ms - 100ms | | 1 MB - 10 MB | 100ms - 1s | | 10 MB - 50 MB | 1s - 5s | | > 50 MB | Blocked |
Built-in File Size Protection
Automatic Size Limits:
- 50 MB maximum for both local files and remote URLs
- Pre-download checking via HTTP Content-Length headers
- Post-download validation for responses without size headers
Size Limit Enforcement:
Local Files:
# File size checked before reading
json_schema({ filePath: "./huge-file.json" })
# → Error: File too large (75MB). This tool is optimized for JSON files under 50MB.Remote URLs:
# Content-Length header checked before download
json_filter({ filePath: "https://api.example.com/massive-dataset" })
# → Error: Response too large (120MB). This tool is optimized for JSON files under 50MB.Size Information:
- All tools display actual file sizes in responses
json_schemashows file size at the top of generated schemasjson_dry_runprovides detailed size breakdowns for optimization- Clear error messages show actual vs. maximum allowed sizes
Performance Recommendations
For Best Performance:
- Use
json_dry_runfirst to analyze large files - Filter data with
json_filterbefore generating schemas - Limit shape specifications to essential fields only
- Process large datasets in smaller chunks when possible
Remote Data Capabilities
Supported Sources
- Public APIs - REST endpoints returning JSON data
- Static JSON files - JSON files hosted on web servers
- Local development -
http://localhostendpoints during development
Content Size Management
- Automatic detection - Checks Content-Length headers before download
- Memory protection - Prevents downloading files larger than 50MB
- Progress indication - Clear error messages showing actual vs. maximum size
- Streaming safety - Validates content size after download for headers without Content-Length
Common Use Cases
For LLMs:
- Filter large API responses to extract only relevant data
- Generate schemas from public API endpoints for integration
- Process configuration files from remote sources
- Analyze sample data from documentation URLs
For Development:
- Extract TypeScript types from API documentation examples
- Filter test data from development APIs
- Process JSON configurations from remote repositories
- Analyze third-party API response structures
Example Workflow:
- LLM calls an API and gets a large response
- Uses
json_filterto extract only needed fields - Processes clean, relevant data without noise
- Generates schemas with
json_schemafor type safety
