Audit how websites respond to AI, search engine, and SEO bot user agents β in bulk
Enter URLs or domains, one per line. We'll check how each responds to selected bot user agents.
| Code | Meaning | Severity | What to Check / What to Do |
|---|---|---|---|
| 200 | Allowed | β Good | Bot can access this URL normally. No action needed. |
| 301 / 302 | Redirect | βΉ Info | URL redirects elsewhere. Common for HTTPβHTTPS or wwwβnon-www. Most bots follow redirects β check the destination URL still returns 200. Excessive redirect chains slow crawling. |
| 403 | Forbidden / Blocked | β High | Server is actively rejecting this bot's User-Agent. Check: (1) robots.txt Disallow rules, (2) Cloudflare WAF/Bot Management settings, (3) .htaccess or nginx UA blocking rules, (4) CDN security configs. If blocking a search engine bot this is a critical SEO issue. |
| 404 | Not Found | β Medium | URL doesn't exist. Verify the URL is correct. If it existed before, check for redirects or content removal. Bots will deindex 404 pages over time. |
| 429 | Rate Limited | β Medium | Too many requests too fast. Increase the delay setting in this tool. For the real bot, review crawl rate settings in Google Search Console or the bot's dashboard. |
| 451 | Legal Block | β Medium | Content blocked for legal reasons (GDPR, CCPA, court order). Usually geo-restricted. Check if this is intentional β if not, review your geo-blocking rules or CDN configuration. |
| 500 / 502 / 503 | Server Error | β High | Server-side failure not specific to the bot. Check server health, error logs, memory/CPU usage. Bots may back off and retry later. Persistent 5xx causes deindexing. |
| TIMEOUT | No Response | β High | Connection opened but no response within 10s. Could be: geo/IP blocking at network level, DDoS protection dropping the connection, very slow server, or firewall silently dropping bot traffic. |
| ERR | Network Error | π¨ Check | Complete connection failure β DNS didn't resolve, server unreachable, or deep firewall block. Verify the domain is live and DNS is propagated. This tool uses a proxy so some networks may cause false positives. |
/admin/) never the whole siterobots.txt for Disallow: / under User-agent: * or User-agent: Googlebotrobots.txt: User-agent: bingbot + Disallow: /robots.txt for Bingbot-specific Disallow rulesbingbotrobots.txt: User-agent: Baiduspider + Disallow: /Baiduspiderrobots.txt for User-agent: Baiduspider Disallowrobots.txt: User-agent: Naverbot + Disallow: /User-agent: GPTBot + Disallow: / in robots.txtUser-agent: ChatGPT-User + Disallow: / in robots.txt<meta name="robots" content="noai, noimageai"> to block image trainingrobots.txt for GPTBot/ChatGPT-User Disallow entriesUser-agent: OAI-SearchBot + Disallow: / in robots.txtrobots.txt β a User-agent: * Disallow will catch this botUser-agent: Google-Extended + Disallow: / in robots.txtUser-agent: Google-Extended Disallow from robots.txtUser-agent: ClaudeBot + Disallow: / in robots.txtUser-agent: * blocksBytespiderUser-agent: CCBot + Disallow: / in robots.txtUser-agent: Amazonbot + Disallow: / in robots.txtUser-agent: cohere-ai + Disallow: / in robots.txtUser-agent: * DisallowUser-agent Disallow rules in robots.txt for eachDiffbotUser-agent: Diffbot + Disallow: / in robots.txtrobots.txt β must NOT have Disallow: / under User-agent: * or User-agent: GooglebotBotBlocker or UA-based deny rulesrobots.txt:User-agent: GPTBot Disallow: /User-agent: ChatGPT-User Disallow: /User-agent: Google-Extended Disallow: /User-agent: ClaudeBot Disallow: /User-agent: CCBot Disallow: /<meta name="robots" content="noai, noimageai">Disallow: / = entire site blocked for that botDisallow: /wp-admin/ = only that path blockedAllow: / overrides a broader Disallow* (wildcard) rules