Page Indexability Checker
Check if a URL can be indexed by Google. Detects 'noindex' tags, robots.txt blocks, and canonical issues.
Why check indexability?
- Ensure your new content can actually appear in Google
- Identify accidental 'noindex' tags left after development
- Verify robots.txt isn't blocking important sections
- Debug why a page has disappeared from search results
- Check if Canonical tags are pointing to the wrong URL
Frequently Asked Questions
What is the difference between 'noindex' and 'robots.txt check'?
'noindex' is a command on the page itself telling Google 'do not index this specific page'. Robots.txt is a gatekeeper that says 'do not even enter/crawl this section'. Google cannot see the 'noindex' tag if it is blocked by robots.txt.
What happens if a Canonical URL is different?
If the canonical tag points to a different URL (A points to B), Google will likely ignore page A and only index page B. This is normal for duplicate content but bad if unintentional.
What is an X-Robots-Tag?
It is an HTTP header sent by the server (not visible in HTML source code) that can control indexing just like a meta tag. It is often used for non-HTML files like PDFs or images.