Fix guide

Fix Pages Blocked by Robots.txt

A robots blocker finding means public pages or discovery files that should be crawlable are disallowed or hidden by policy.

Read-only public scan. No login, no crawler install, and no private URLs are fetched.

Surface
Fix guide
Scope
Public web evidence
Auth
None required
Schema
TechArticle

Answer first

What the issue means

robots.txt returned rules that block sampled public pages, sitemap URLs, llms.txt, or important AI/search crawler user-agents.

Detail 01

Exact fix

Keep private paths blocked, but allow the homepage, canonical content, sitemap.xml, robots.txt, llms.txt, and public assets needed for previews.

Detail 02

Verify

Test the exact user-agent rule, fetch the blocked URL, check sitemap coverage, and run a fresh crawler-access scan.

FAQ

Common questions

Can blocked pages still be indexed?
Sometimes yes if other pages link to them, but crawlers cannot fetch the content to evaluate or refresh it.
Should I allow every bot?
No. Decide your policy intentionally. The issue is accidental blocking or ambiguous policy around pages you want discoverable.