npm package discovery and stats viewer.

Discover Tips

  • General search

    [free text search, go nuts!]

  • Package details

    pkg:[package-name]

  • User packages

    @[username]

Sponsor

Optimize Toolset

I’ve always been into building performant and accessible sites, but lately I’ve been taking it extremely seriously. So much so that I’ve been building a tool to help me optimize and monitor the sites that I build to make sure that I’m making an attempt to offer the best experience to those who visit them. If you’re into performant, accessible and SEO friendly sites, you might like it too! You can check it out at Optimize Toolset.

About

Hi, 👋, I’m Ryan Hefner  and I built this site for me, and you! The goal of this site was to provide an easy way for me to check the stats on my npm packages, both for prioritizing issues and updates, and to give me a little kick in the pants to keep up on stuff.

As I was building it, I realized that I was actually using the tool to build the tool, and figured I might as well put this out there and hopefully others will find it to be a fast and useful way to search and browse npm packages as I have.

If you’re interested in other things I’m working on, follow me on Twitter or check out the open source projects I’ve been publishing on GitHub.

I am also working on a Twitter bot for this site to tweet the most popular, newest, random packages from npm. Please follow that account now and it will start sending out packages soon–ish.

Open Software & Tools

This site wouldn’t be possible without the immense generosity and tireless efforts from the people who make contributions to the world and share their work via open source initiatives. Thank you 🙏

© 2026 – Pkg Stats / Ryan Hefner

n8n-nodes-anakin-scraper

v1.1.0

Published

n8n node for Anakin API - web scraping, AI search, and intelligent data extraction

Downloads

99

Readme

n8n-nodes-anakin-scraper

This is an n8n community node that lets you use the Anakin API in your n8n workflows.

Anakin provides powerful web scraping, AI-powered search, and intelligent data extraction capabilities. This node handles all the complexity of job submission and polling automatically.

Features

  • 🔐 Simple authentication with API key
  • 🌐 Web Scraping: Scrape any website and extract structured data
  • 🔍 AI Search: Perform intelligent searches powered by Perplexity AI
  • 🤖 Agentic Search: Advanced multi-stage pipeline that searches, scrapes, and extracts structured data automatically
  • ⏳ Automatic polling for async operations
  • 🎯 Configurable polling intervals and timeouts
  • 🌍 Support for country-specific proxy routing
  • ♻️ Cache control with force fresh option

Installation

Community Nodes (Recommended)

  1. Go to Settings > Community Nodes in your n8n instance
  2. Select Install
  3. Enter n8n-nodes-anakin-scraper in the Package name field
  4. Click Install

Manual Installation

Navigate to your n8n installation folder and run:

npm install n8n-nodes-anakin-scraper

Then restart n8n.

Setup

1. Configure Credentials

Before using the Anakin node, you need to set up your API credentials:

  1. In n8n, go to Credentials > New
  2. Search for Anakin API
  3. Fill in:
    • API Key: Your Anakin Scraper API authentication token
    • Base URL: The API endpoint (default: https://api.anakin.io)
  4. Click Save

2. Use in Workflow

  1. Add the Anakin node to your workflow
  2. Connect it to your trigger or previous node
  3. Select your credentials
  4. Enter the URL you want to scrape
  5. (Optional) Configure additional options

Usage

The Anakin node supports three powerful operations:

1. Scrape URL

Extract content and structured data from any website.

Trigger → Anakin (Scrape URL) → Process Data

Configuration:

  • URL: https://example.com/product-page
  • Country Code: us (optional)
  • Force Fresh: false (optional)
  • Max Wait Time: 300 seconds (optional)
  • Poll Interval: 3 seconds (optional)

Output:

{
  "success": true,
  "operation": "scrapeUrl",
  "request_id": "req_123456",
  "url": "https://example.com/product-page",
  "status": "completed",
  "html": "...",
  "markdown": "...",
  "generatedJson": {
    // Structured data extracted from the page
  }
}

2. Search

Perform AI-powered searches using Perplexity AI. Get instant answers with citations.

Trigger → Anakin (Search) → Process Results

Configuration:

  • Search Query: What are the latest trends in AI?
  • Max Results: 5 (optional, default: 5)

Output:

{
  "success": true,
  "operation": "search",
  "query": "What are the latest trends in AI?",
  "answer": "Based on recent developments...",
  "results": [
    {
      "title": "AI Trends 2026",
      "url": "https://example.com/ai-trends",
      "content": "Summary of the article...",
      "score": 0.95
    }
  ],
  "count": 5
}

Use Cases:

  • Research and fact-checking
  • Competitive intelligence
  • Content research
  • Real-time information gathering

3. Agentic Search

Advanced multi-stage AI pipeline that automatically:

  1. Searches for relevant information
  2. Identifies and scrapes citation sources
  3. Extracts structured data using AI
  4. Generates a comprehensive summary
Trigger → Anakin (Agentic Search) → Process Structured Data

Configuration:

  • Search Prompt: Find the pricing plans for top 5 CRM software
  • Use Browser: true (optional, more reliable)
  • Max Wait Time: 600 seconds (optional)
  • Poll Interval: 5 seconds (optional)

Output:

{
  "success": true,
  "operation": "agenticSearch",
  "job_id": "job_789",
  "status": "completed",
  "query": "Find the pricing plans for top 5 CRM software",
  "perplexity_answer": "Here are the top CRM solutions...",
  "citations": [
    {"url": "https://salesforce.com/pricing", "title": "Salesforce Pricing", "source_index": 0}
  ],
  "scraped_data": [
    {
      "source_url": "https://salesforce.com/pricing",
      "source_index": 0,
      "data": {
        "plans": [...],
        "features": [...]
      }
    }
  ],
  "chatgpt_schema": {
    "type": "object",
    "properties": {...}
  },
  "chatgpt_structured_data": {
    "crm_platforms": [...]
  },
  "chatgpt_summary": "Comprehensive analysis of CRM pricing..."
}

Use Cases:

  • Market research with structured data
  • Competitive analysis
  • Lead generation with enriched data
  • Automated data collection for reports

Configuration Options by Operation

Scrape URL Options

| Option | Description | Default | |--------|-------------|---------| | URL | The website URL to scrape | Required | | Max Wait Time | Maximum seconds to wait for completion | 300 | | Poll Interval | Seconds between status checks | 3 | | Country Code | Proxy country code (e.g., us, uk, de) | us | | Force Fresh | Bypass cache and force fresh scrape | false |

Search Options

| Option | Description | Default | |--------|-------------|---------| | Search Query | The question or query to search | Required | | Max Results | Maximum number of results to return | 5 |

Agentic Search Options

| Option | Description | Default | |--------|-------------|---------| | Search Prompt | The search prompt for analysis | Required | | Use Browser | Use browser for scraping (more reliable) | true | | Max Wait Time | Maximum seconds to wait for completion | 600 | | Poll Interval | Seconds between status checks | 5 |

How It Works

Scrape URL & Agentic Search (Async)

  1. Submit: The node submits your request to the Anakin API
  2. Poll: Automatically checks the job status every few seconds
  3. Return: Once complete, returns the data to your workflow

Search (Synchronous)

  1. Submit: The node sends your query to the Anakin API
  2. Return: Immediately returns search results with answers and citations

The node handles all the complexity of:

  • Job submission and request management
  • Automatic polling for async operations
  • Intelligent error handling
  • Timeout management
  • Response parsing and formatting

Error Handling

The node will throw an error if:

  • The scraping job fails
  • The job doesn't complete within the max wait time
  • The API returns an error

You can enable Continue on Fail in the node settings to handle errors gracefully.

API Endpoints Used

Scrape URL

  • POST /v1/request - Submit scraping job
  • GET /v1/request/{id} - Check job status

Search

  • POST /v1/search - Perform AI search (synchronous)

Agentic Search

  • POST /v1/agentic-search - Submit agentic search job
  • GET /v1/agentic-search/{jobId} - Check agentic search status

Development

Prerequisites

  • Node.js >= 16
  • n8n installed locally

Setup

# Clone the repository
git clone https://github.com/Anakin-Inc/blueprint-scribe-35.git
cd blueprint-scribe-35/external-integrations/n8n

# Install dependencies
npm install

# Build the node
npm run build

# Link for local development
npm link
cd ~/.n8n/custom
npm link n8n-nodes-anakin-scraper

Project Structure

n8n-nodes-anakin-scraper/
├── credentials/
│   └── AnakinScraperApi.credentials.ts
├── nodes/
│   └── AnakinScraper/
│       └── AnakinScraper.node.ts
├── package.json
└── README.md

Support

For issues, questions, or contributions:

License

MIT

Changelog

1.1.0

  • Added AI-powered Search operation (Perplexity integration)
  • Added Agentic Search operation (multi-stage AI pipeline)
  • Improved error handling and logging
  • Updated node display name to "Anakin"

1.0.0

  • Initial release
  • URL scraping with automatic polling
  • Configurable polling intervals and timeouts
  • Country-specific proxy support
  • Cache control options