Index Coverage is a status report within Google Search Console that identifies which pages of a website Google has successfully crawled and indexed. It provides feedback on technical issues that may prevent specific URLs from appearing in search results. Monitoring this report ensures that your important content is eligible to drive organic traffic.
What is Index Coverage?
The Index Coverage report (often called the Page Indexing report) shows the indexing state of all URLs Google has visited or tried to visit on your property. It categorizes pages into four main statuses: Valid, Valid with Warnings, Excluded, and Error. This feedback helps webmasters understand how Google perceives their site's technical structure.
While Google states that sites with fewer than 500 pages likely do not need this report, some practitioners argue it remains essential for any business relying on search traffic because it is more reliable than using the site: search operator.
Why Index Coverage matters
- Traffic Visibility: Only indexed pages can appear in Google Search and Discover results.
- Technical Troubleshooting: Identifies server errors, redirect loops, and crawl blocks that waste crawl budget.
- Content Optimization: Confirms whether Google is prioritizing your preferred (canonical) pages over duplicates.
- Growth Tracking: A healthy site should show a gradually increasing count of indexed pages as it grows.
- Error Prevention: Alerts you to sudden drops in indexed pages, which often signal significant site-wide issues.
How Index Coverage works
Indexing is a multi-step process that moves from discovery to eligibility for ranking:
- Discovery: Google finds URLs by following links from other pages or processing submitted XML sitemaps.
- Crawling: During this phase, Googlebot requests the URL and gathers information. Google follows five redirects per crawl attempt before stopping.
- Indexing: The indexer tries to make sense of the page content, determines its relevance, and adds it to the Google index.
Once indexed, a page is eligible to appear in search results, though appearance is not guaranteed as results are customized based on user history and location.
Status types
The report groups URLs into four categories based on their health:
| Status | Meaning | Action Needed |
|---|---|---|
| Valid | Pages are successfully indexed. | None, or add to sitemap if missing. |
| Valid with Warnings | Indexed, but have issues like being blocked by robots.txt. | Review to ensure robots.txt isn't blocking valid content. |
| Excluded | Purposefully not indexed (e.g., duplicates, redirects). | Verify that important pages aren't accidentally excluded. |
| Error | Cannot be indexed due to a critical problem (e.g., 404 or 5xx). | Immediate troubleshooting and fixing required. |
Best practices
- Focus on canonicals: Your goal is to get the canonical version of every important page indexed. Do not expect 100% coverage of every URL on your site, as duplicates should be excluded.
- Monitor spikes and drops: Investigating sudden changes in total indexed pages. Spikes in errors often stem from template changes or robots.txt misconfigurations.
- Use sitemaps for granularity: Create separate sitemaps for different folders or content types to pinpoint where indexing issues occur.
- Fix "Website" sources first: The report identifies the source of issues as either "Google" or "Website." Focus your efforts on fixing issues labeled as "Website."
- Validate fixes promptly: Once you fix an issue, use the Validate Fix button. Validation typically takes up to about two weeks, though it can take longer.
Common mistakes
- Mistake: Using robots.txt to "noindex" a page. Fix: Robots.txt only stops crawling. To stop indexing, you must remove the robots.txt block and use a proper "noindex" tag or HTTP header.
- Mistake: Expecting immediate indexing for new content. Fix: It can take a few days or weeks for Google to index new pages. Use the URL Inspection tool to request indexing for priority pages.
- Mistake: Ignoring "Excluded" pages. Fix: While many exclusions are normal (like redirects), important pages may be excluded due to "Soft 404" errors or being "Discovered but not indexed."
- Mistake: Relying on the 1,000-row sample. Fix: The example list in the report is limited to 1,000 items. If you have more errors, you must fix the known samples and validate to see the remaining URLs.
Examples
- Redirect Error: You have a redirect chain that is too long or a redirect loop that prevents Googlebot from reaching the target page.
- Soft 404: A page returns a 200 "Success" code but displays a "Page Not Found" message to users.
- Indexed, though blocked: Google found a link to a page blocked by your robots.txt. It indexed the URL based on external information, but cannot "see" the content on the page.
Index Coverage vs URL Inspection
| Feature | Index Coverage Report | URL Inspection Tool |
|---|---|---|
| Scale | Bulk overview of the whole site. | Single URL deep dive. |
| Data Type | Historical data from the last crawl. | Real-time "Live Test" available. |
| Purpose | Identifying site-wide trends and errors. | Debugging specific pages and requesting indexing. |
| Visibility | Shows a sample of up to 1,000 URLs. | Shows exact status for any known URL. |
FAQ
Why are so many of my pages "Excluded"?
This is often normal. Most excluded URLs are duplicates, pages with redirects, or URLs intentionally blocked by "noindex" tags. Large sites often have more excluded URLs than indexed ones due to parameters (like ?sort=price) that Google correctly identifies as non-canonical duplicates.
How do I get Google to recrawl a fixed page? You can use the URL Inspection tool to ask Google to recrawl individual pages. For bulk fixes, update your site and click "Validate Fix" in the Index Coverage report. Google will naturally recrawl pages over time based on how often they change.
What is "Crawled - currently not indexed"? This means Googlebot successfully visited the page but has not added it to the index. It may be due to low content quality, thin content, or because the page was only recently crawled and is still being processed.
What happened to "Crawl Anomaly"? Following the January 2021 update, Google removed the generic "Crawl Anomaly" type. These errors are now mapped to more specific categories like server errors or redirect issues to make them easier to fix.
Can I hide a URL from search results temporarily? Yes, the Google Page Removal tool can hide URLs for a brief period. Hidden URLs are removed for 90 days before they may reappear. You should use "noindex" tags for a permanent solution.