nuxt3-bot-handler
v1.0.8-beta
Published
🛡️ Nuxt 3 middleware to block suspicious bots, protect SEO crawlers with reverse DNS checks, and enforce User-Agent rules.
Maintainers
Readme
nuxt3-bot-handler
🛡️ A Nuxt 3 server middleware to detect and block suspicious bots, protect SEO integrity, and allow only verified crawlers using reverse DNS validation and user-agent structure analysis.
✨ Features
- Detects malformed or spoofed user-agents
- DNS reverse lookup verification for SEO crawlers
- Blocks common scraping tools (curl, wget, headlesschrome, etc.)
- Whitelists official crawlers (Googlebot, Bingbot, Twitterbot, Applebot, etc.)
- Passes uptime checkers like
Uptime Kumasafely - Easy plug-and-play in any Nuxt 3 project
- Customizable verbosity with options
📦 Installation
With npm
npm install nuxt3-bot-handlerWith pnpm
pnpm add nuxt3-bot-handler🧩 Usage
In your Nuxt 3 project, add the middleware like this:
// server/middleware/bot-handler.ts
import { createBotHandler } from 'nuxt3-bot-handler'
export default createBotHandler({ verbose: true })Or for minimal logging:
export default createBotHandler({ verbose: false })That's it — Nuxt will automatically run this middleware for every incoming request.
🔍 How It Works
This middleware performs the following checks:
User-Agent Validation
Blocks missing, too short, or generic user-agents (like "test", "curl", etc.)Suspicious Pattern Detection
Matches against a list of known bot/scraper patterns (axios,wget,headlesschrome, etc.)DNS Reverse Lookup for SEO bots
Verifies that the IP address belongs to the official domain of bots (e.g., Googlebot must resolve to *.googlebot.com) using dns.reverse(). If the DNS reverse lookup fails due to network issues or unresolvable hostnames, the request is blocked to avoid spoofing. However, if the error is caused by unsupported functionality (e.g., "Not implemented: cares.ChannelWrap.prototype.getHostByAddr"), the lookup is skipped, and a warning is logged (if verbose mode is enabled), without blocking the request. This prevents false positives in restricted environments such as edge runtimes or some serverless deployments.Bypasses for Facebook and Meta IPs
Allows Facebook crawlers with specific IPv4/IPv6 prefixes even without reverse DNSStructural Checks on User-Agent
Denies clients with flat or malformed User-Agent strings, unless explicitly allowlistedVerbose Option for Logging
Toggle detailed console logging using theverbose: true|falseoption
✅ Whitelisted Crawlers
The middleware allows through these bots after DNS check:
- Googlebot
- AdsBot-Google
- Bingbot
- DuckDuckBot
- Yahoo Slurp
- YandexBot
- Applebot
- SemrushBot
- Screaming Frog SEO Spider
- Twitterbot
- facebot / facebookexternalhit / meta-externalagent
- uptime-kuma
- Cookiebot
- Greenflare
🧪 Testing
To test locally:
npx nuxi dev
curl -A "curl/7.77.0" http://localhost:3000Should return 403 Forbidden.
📜 License
MIT © Lorenzo Furno
🤝 Contributing
Pull requests are welcome! If you have suggestions or want to help support more bots, open an issue or PR.
