npm package discovery and stats viewer.

Discover Tips

  • General search

    [free text search, go nuts!]

  • Package details

    pkg:[package-name]

  • User packages

    @[username]

Sponsor

Optimize Toolset

I’ve always been into building performant and accessible sites, but lately I’ve been taking it extremely seriously. So much so that I’ve been building a tool to help me optimize and monitor the sites that I build to make sure that I’m making an attempt to offer the best experience to those who visit them. If you’re into performant, accessible and SEO friendly sites, you might like it too! You can check it out at Optimize Toolset.

About

Hi, 👋, I’m Ryan Hefner  and I built this site for me, and you! The goal of this site was to provide an easy way for me to check the stats on my npm packages, both for prioritizing issues and updates, and to give me a little kick in the pants to keep up on stuff.

As I was building it, I realized that I was actually using the tool to build the tool, and figured I might as well put this out there and hopefully others will find it to be a fast and useful way to search and browse npm packages as I have.

If you’re interested in other things I’m working on, follow me on Twitter or check out the open source projects I’ve been publishing on GitHub.

I am also working on a Twitter bot for this site to tweet the most popular, newest, random packages from npm. Please follow that account now and it will start sending out packages soon–ish.

Open Software & Tools

This site wouldn’t be possible without the immense generosity and tireless efforts from the people who make contributions to the world and share their work via open source initiatives. Thank you 🙏

© 2025 – Pkg Stats / Ryan Hefner

@mcpflow.io/mcp-search1api-mcp-server

v1.0.1

Published

A Model Context Protocol (MCP) server that provides search and crawl functionality using Search1API

Readme

Search1API MCP Server

此包由 MCPFlow 打包并发布到npm仓库。

安装与使用

直接使用npx运行:

npx @mcpflow.io/mcp-search1api-mcp-server

或者先安装后使用:

# 安装
npm install @mcpflow.io/mcp-search1api-mcp-server

# 使用
npx @mcpflow.io/mcp-search1api-mcp-server

使用方法

请参考原始仓库的使用说明。

工具函数

1

参数:

原始信息

原始README

Search1API MCP Server

中文文档

A Model Context Protocol (MCP) server that provides search and crawl functionality using Search1API.

https://github.com/user-attachments/assets/58bc98ae-3b6b-442c-a7fc-010508b5f028

More discussions and updates, please follow our official X, or join the official discord

Features

  • Web search functionality
  • News search functionality
  • Web page content extraction
  • Website sitemap extraction
  • Deep thinking and complex problem solving with DeepSeek R1
  • Seamless integration with Claude Desktop, Cursor, Windsurf, Cline and other MCP clients

Tools

1. Search Tool

  • Name: search
  • Description: Search the web using Search1API
  • Parameters:
    • query (required): Search query in natural language. Be specific and concise for better results
    • max_results (optional, default: 10): Number of results to return
    • search_service (optional, default: "google"): Search service to use (google, bing, duckduckgo, yahoo, x, reddit, github, youtube, arxiv, wechat, bilibili, imdb, wikipedia)
    • crawl_results (optional, default: 0): Number of results to crawl for full webpage content
    • include_sites (optional): List of sites to include in search
    • exclude_sites (optional): List of sites to exclude from search
    • time_range (optional): Time range for search results ("day", "month", "year")

2. News Tool

  • Name: news
  • Description: Search for news articles using Search1API
  • Parameters:
    • query (required): Search query in natural language. Be specific and concise for better results
    • max_results (optional, default: 10): Number of results to return
    • search_service (optional, default: "bing"): Search service to use (google, bing, duckduckgo, yahoo, hackernews)
    • crawl_results (optional, default: 0): Number of results to crawl for full webpage content
    • include_sites (optional): List of sites to include in search
    • exclude_sites (optional): List of sites to exclude from search
    • time_range (optional): Time range for search results ("day", "month", "year")

3. Crawl Tool

  • Name: crawl
  • Description: Extract content from a URL using Search1API
  • Parameters:
    • url (required): URL to crawl

4. Sitemap Tool

  • Name: sitemap
  • Description: Get all related links from a URL
  • Parameters:
    • url (required): URL to get sitemap

5. Reasoning Tool

  • Name: reasoning
  • Description: A tool for deep thinking and complex problem solving with fast deepseek r1 model and web search ability(You can change to any other model in search1api website but the speed is not guaranteed)
  • Parameters:
    • content (required): The question or problem that needs deep thinking

6. Trending Tool

  • Name: trending
  • Description: Get trending topics from popular platforms
  • Parameters:
    • search_service (required): Specify the platform to get trending topics from (github, hackernews)
    • max_results (optional, default: 10): Maximum number of trending items to return

Setup Guide

1. Get Search1API Key

  1. Register at Search1API
  2. Get your api key and 100 free credits

2. Configure

{
  "mcpServers": {
    "search1api": {
      "command": "npx",
      "args": ["-y", "search1api-mcp"],
      "env": {
        "SEARCH1API_KEY": "YOUR_SEARCH1API_KEY"
      }
    }
  }
}

Version History

  • v0.1.8: Added X(Twitter) and Reddit search services
  • v0.1.7: Added Trending tool for GitHub and Hacker News
  • v0.1.6: Added Wikipedia search service
  • v0.1.5: Added new search parameters (include_sites, exclude_sites, time_range) and new search services (arxiv, wechat, bilibili, imdb)
  • v0.1.4: Added reasoning tool with deepseek r1 and updated the Cursor and Windsurf configuration guide
  • v0.1.3: Added news search functionality
  • v0.1.2: Added sitemap functionality
  • v0.1.1: Added web crawling functionality
  • v0.1.0: Initial release with search functionality

License

This project is licensed under the MIT License - see the LICENSE file for details.