Page Indexability Checker

Check if a URL can be indexed by Google. Detects 'noindex' tags, robots.txt blocks, and canonical issues.

Why check indexability?

  • Ensure your new content can actually appear in Google
  • Identify accidental 'noindex' tags left after development
  • Verify robots.txt isn't blocking important sections
  • Debug why a page has disappeared from search results
  • Check if Canonical tags are pointing to the wrong URL

Frequently Asked Questions

What is the difference between 'noindex' and 'robots.txt check'?

'noindex' is a command on the page itself telling Google 'do not index this specific page'. Robots.txt is a gatekeeper that says 'do not even enter/crawl this section'. Google cannot see the 'noindex' tag if it is blocked by robots.txt.

What happens if a Canonical URL is different?

If the canonical tag points to a different URL (A points to B), Google will likely ignore page A and only index page B. This is normal for duplicate content but bad if unintentional.

What is an X-Robots-Tag?

It is an HTTP header sent by the server (not visible in HTML source code) that can control indexing just like a meta tag. It is often used for non-HTML files like PDFs or images.