Page Indexability Checker
Check if a URL can be indexed by Google. Detects 'noindex' tags, robots.txt blocks, and canonical issues.
Why check indexability?
- Ensure your new content can actually appear in Google
- Identify accidental 'noindex' tags left after development
- Verify robots.txt isn't blocking important sections
- Debug why a page has disappeared from search results
- Check if Canonical tags are pointing to the wrong URL
Frequently Asked Questions
What is the difference between 'noindex' and 'robots.txt check'?
'noindex' is a command on the page itself telling Google 'do not index this specific page'. Robots.txt is a gatekeeper that says 'do not even enter/crawl this section'. Google cannot see the 'noindex' tag if it is blocked by robots.txt.
What happens if a Canonical URL is different?
If the canonical tag points to a different URL (A points to B), Google will likely ignore page A and only index page B. This is normal for duplicate content but bad if unintentional.
What is an X-Robots-Tag?
It is an HTTP header sent by the server (not visible in HTML source code) that can control indexing just like a meta tag. It is often used for non-HTML files like PDFs or images.
Results are generated in real-time. For best accuracy, verify critical issues manually.
What this tool checks
- ✓ Robots.txt Disallow Rules
- ✓ Meta Robots 'noindex' Tag
- ✓ X-Robots-Tag HTTP Header
- ✓ Canonical URL Consistency
Common problems this tool finds
- ⚠️ Dev sites left in 'noindex' mode
- ⚠️ Accidental robots.txt blocking entire site
- ⚠️ Canonical pointing to HTTP version
- ⚠️ Conflicting rules (Allowed in robots.txt but noindex in meta)
- ⚠️ Password protected pages (401/403)
How to fix results (Quick Checklist)
- 1.Remove '<meta name="robots" content="noindex">' from the page HTML.
- 2.Remove 'Disallow: /your-page/' from the robots.txt file.
- 3.Ensure the canonical tag points to the current URL (self-referencing).
- 4.Request indexing in Google Search Console after fixing blocks.