Fix Pages Blocked by Robots.txt
A robots blocker finding means public pages or discovery files that should be crawlable are disallowed or hidden by policy.
- Surface
- Fix guide
- Scope
- Public web evidence
- Auth
- None required
- Schema
- TechArticle
Answer first
What the issue means
robots.txt returned rules that block sampled public pages, sitemap URLs, llms.txt, or important AI/search crawler user-agents.
Detail 01
Exact fix
Keep private paths blocked, but allow the homepage, canonical content, sitemap.xml, robots.txt, llms.txt, and public assets needed for previews.
Detail 02
Verify
Test the exact user-agent rule, fetch the blocked URL, check sitemap coverage, and run a fresh crawler-access scan.
FAQ
Common questions
- Can blocked pages still be indexed?
- Sometimes yes if other pages link to them, but crawlers cannot fetch the content to evaluate or refresh it.
- Should I allow every bot?
- No. Decide your policy intentionally. The issue is accidental blocking or ambiguous policy around pages you want discoverable.